00:00:00.001 Started by upstream project "autotest-per-patch" build number 131276 00:00:00.001 originally caused by: 00:00:00.001 Started by user sys_sgci 00:00:00.041 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-phy.groovy 00:00:00.041 The recommended git tool is: git 00:00:00.042 using credential 00000000-0000-0000-0000-000000000002 00:00:00.043 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.070 Fetching changes from the remote Git repository 00:00:00.072 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.124 Using shallow fetch with depth 1 00:00:00.124 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.124 > git --version # timeout=10 00:00:00.194 > git --version # 'git version 2.39.2' 00:00:00.194 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.271 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.271 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:03.735 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:03.750 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:03.762 Checking out Revision 58e4f482292076ec19d68e6712473e60ef956aed (FETCH_HEAD) 00:00:03.762 > git config core.sparsecheckout # timeout=10 00:00:03.776 > git read-tree -mu HEAD # timeout=10 00:00:03.792 > git checkout -f 58e4f482292076ec19d68e6712473e60ef956aed # timeout=5 00:00:03.812 Commit message: "packer: Fix typo in a package name" 00:00:03.813 > git rev-list --no-walk 58e4f482292076ec19d68e6712473e60ef956aed # timeout=10 00:00:03.926 [Pipeline] Start of Pipeline 00:00:03.940 [Pipeline] library 00:00:03.942 Loading library shm_lib@master 00:00:03.942 Library shm_lib@master is cached. Copying from home. 00:00:03.958 [Pipeline] node 00:00:03.969 Running on CYP13 in /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:03.971 [Pipeline] { 00:00:03.984 [Pipeline] catchError 00:00:03.986 [Pipeline] { 00:00:04.001 [Pipeline] wrap 00:00:04.011 [Pipeline] { 00:00:04.021 [Pipeline] stage 00:00:04.023 [Pipeline] { (Prologue) 00:00:04.252 [Pipeline] sh 00:00:04.543 + logger -p user.info -t JENKINS-CI 00:00:04.563 [Pipeline] echo 00:00:04.565 Node: CYP13 00:00:04.571 [Pipeline] sh 00:00:04.875 [Pipeline] setCustomBuildProperty 00:00:04.885 [Pipeline] echo 00:00:04.886 Cleanup processes 00:00:04.890 [Pipeline] sh 00:00:05.175 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:05.175 3910376 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:05.187 [Pipeline] sh 00:00:05.474 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:05.474 ++ grep -v 'sudo pgrep' 00:00:05.474 ++ awk '{print $1}' 00:00:05.474 + sudo kill -9 00:00:05.474 + true 00:00:05.489 [Pipeline] cleanWs 00:00:05.499 [WS-CLEANUP] Deleting project workspace... 00:00:05.499 [WS-CLEANUP] Deferred wipeout is used... 00:00:05.507 [WS-CLEANUP] done 00:00:05.510 [Pipeline] setCustomBuildProperty 00:00:05.519 [Pipeline] sh 00:00:05.802 + sudo git config --global --replace-all safe.directory '*' 00:00:05.881 [Pipeline] httpRequest 00:00:06.275 [Pipeline] echo 00:00:06.277 Sorcerer 10.211.164.101 is alive 00:00:06.286 [Pipeline] retry 00:00:06.288 [Pipeline] { 00:00:06.302 [Pipeline] httpRequest 00:00:06.307 HttpMethod: GET 00:00:06.307 URL: http://10.211.164.101/packages/jbp_58e4f482292076ec19d68e6712473e60ef956aed.tar.gz 00:00:06.308 Sending request to url: http://10.211.164.101/packages/jbp_58e4f482292076ec19d68e6712473e60ef956aed.tar.gz 00:00:06.311 Response Code: HTTP/1.1 200 OK 00:00:06.311 Success: Status code 200 is in the accepted range: 200,404 00:00:06.312 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/jbp_58e4f482292076ec19d68e6712473e60ef956aed.tar.gz 00:00:06.457 [Pipeline] } 00:00:06.471 [Pipeline] // retry 00:00:06.476 [Pipeline] sh 00:00:06.768 + tar --no-same-owner -xf jbp_58e4f482292076ec19d68e6712473e60ef956aed.tar.gz 00:00:06.782 [Pipeline] httpRequest 00:00:07.231 [Pipeline] echo 00:00:07.233 Sorcerer 10.211.164.101 is alive 00:00:07.241 [Pipeline] retry 00:00:07.243 [Pipeline] { 00:00:07.254 [Pipeline] httpRequest 00:00:07.259 HttpMethod: GET 00:00:07.260 URL: http://10.211.164.101/packages/spdk_18d1cad7be4b3bcc316224e1f0669c83846fcc94.tar.gz 00:00:07.261 Sending request to url: http://10.211.164.101/packages/spdk_18d1cad7be4b3bcc316224e1f0669c83846fcc94.tar.gz 00:00:07.270 Response Code: HTTP/1.1 200 OK 00:00:07.270 Success: Status code 200 is in the accepted range: 200,404 00:00:07.270 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk_18d1cad7be4b3bcc316224e1f0669c83846fcc94.tar.gz 00:00:22.507 [Pipeline] } 00:00:22.525 [Pipeline] // retry 00:00:22.534 [Pipeline] sh 00:00:22.827 + tar --no-same-owner -xf spdk_18d1cad7be4b3bcc316224e1f0669c83846fcc94.tar.gz 00:00:26.171 [Pipeline] sh 00:00:26.456 + git -C spdk log --oneline -n5 00:00:26.456 18d1cad7b nvme/rdma: Register UMR per IO request 00:00:26.456 88ed43660 accel/mlx5: Support mkey registration 00:00:26.456 b1174c3d3 accel/mlx5: Create pool of UMRs 00:00:26.456 652e05bb1 lib/mlx5: API to configure UMR 00:00:26.456 36b1ba373 accel/mlx5: Register post_poller handler 00:00:26.468 [Pipeline] } 00:00:26.481 [Pipeline] // stage 00:00:26.489 [Pipeline] stage 00:00:26.491 [Pipeline] { (Prepare) 00:00:26.507 [Pipeline] writeFile 00:00:26.522 [Pipeline] sh 00:00:26.809 + logger -p user.info -t JENKINS-CI 00:00:26.822 [Pipeline] sh 00:00:27.111 + logger -p user.info -t JENKINS-CI 00:00:27.124 [Pipeline] sh 00:00:27.414 + cat autorun-spdk.conf 00:00:27.415 SPDK_RUN_FUNCTIONAL_TEST=1 00:00:27.415 SPDK_TEST_NVMF=1 00:00:27.415 SPDK_TEST_NVME_CLI=1 00:00:27.415 SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:27.415 SPDK_TEST_NVMF_NICS=e810 00:00:27.415 SPDK_TEST_VFIOUSER=1 00:00:27.415 SPDK_RUN_UBSAN=1 00:00:27.415 NET_TYPE=phy 00:00:27.424 RUN_NIGHTLY=0 00:00:27.428 [Pipeline] readFile 00:00:27.451 [Pipeline] withEnv 00:00:27.452 [Pipeline] { 00:00:27.464 [Pipeline] sh 00:00:27.754 + set -ex 00:00:27.754 + [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf ]] 00:00:27.754 + source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:00:27.754 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:00:27.754 ++ SPDK_TEST_NVMF=1 00:00:27.754 ++ SPDK_TEST_NVME_CLI=1 00:00:27.754 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:27.754 ++ SPDK_TEST_NVMF_NICS=e810 00:00:27.754 ++ SPDK_TEST_VFIOUSER=1 00:00:27.754 ++ SPDK_RUN_UBSAN=1 00:00:27.754 ++ NET_TYPE=phy 00:00:27.754 ++ RUN_NIGHTLY=0 00:00:27.754 + case $SPDK_TEST_NVMF_NICS in 00:00:27.754 + DRIVERS=ice 00:00:27.754 + [[ tcp == \r\d\m\a ]] 00:00:27.754 + [[ -n ice ]] 00:00:27.754 + sudo rmmod mlx4_ib mlx5_ib irdma i40iw iw_cxgb4 00:00:27.754 rmmod: ERROR: Module mlx4_ib is not currently loaded 00:00:35.905 rmmod: ERROR: Module irdma is not currently loaded 00:00:35.905 rmmod: ERROR: Module i40iw is not currently loaded 00:00:35.905 rmmod: ERROR: Module iw_cxgb4 is not currently loaded 00:00:35.905 + true 00:00:35.905 + for D in $DRIVERS 00:00:35.905 + sudo modprobe ice 00:00:35.905 + exit 0 00:00:35.916 [Pipeline] } 00:00:35.929 [Pipeline] // withEnv 00:00:35.934 [Pipeline] } 00:00:35.947 [Pipeline] // stage 00:00:35.955 [Pipeline] catchError 00:00:35.957 [Pipeline] { 00:00:35.970 [Pipeline] timeout 00:00:35.970 Timeout set to expire in 1 hr 0 min 00:00:35.972 [Pipeline] { 00:00:35.984 [Pipeline] stage 00:00:35.986 [Pipeline] { (Tests) 00:00:35.999 [Pipeline] sh 00:00:36.289 + jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:36.289 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:36.289 + DIR_ROOT=/var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:36.289 + [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest ]] 00:00:36.289 + DIR_SPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:36.289 + DIR_OUTPUT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:00:36.289 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk ]] 00:00:36.289 + [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:00:36.289 + mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:00:36.289 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:00:36.289 + [[ nvmf-tcp-phy-autotest == pkgdep-* ]] 00:00:36.289 + cd /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:36.289 + source /etc/os-release 00:00:36.289 ++ NAME='Fedora Linux' 00:00:36.289 ++ VERSION='39 (Cloud Edition)' 00:00:36.289 ++ ID=fedora 00:00:36.289 ++ VERSION_ID=39 00:00:36.289 ++ VERSION_CODENAME= 00:00:36.289 ++ PLATFORM_ID=platform:f39 00:00:36.289 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:00:36.289 ++ ANSI_COLOR='0;38;2;60;110;180' 00:00:36.289 ++ LOGO=fedora-logo-icon 00:00:36.289 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:00:36.289 ++ HOME_URL=https://fedoraproject.org/ 00:00:36.289 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:00:36.289 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:00:36.289 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:00:36.289 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:00:36.289 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:00:36.289 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:00:36.289 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:00:36.289 ++ SUPPORT_END=2024-11-12 00:00:36.289 ++ VARIANT='Cloud Edition' 00:00:36.289 ++ VARIANT_ID=cloud 00:00:36.289 + uname -a 00:00:36.289 Linux spdk-cyp-13 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:00:36.289 + sudo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:00:39.595 Hugepages 00:00:39.595 node hugesize free / total 00:00:39.595 node0 1048576kB 0 / 0 00:00:39.595 node0 2048kB 0 / 0 00:00:39.595 node1 1048576kB 0 / 0 00:00:39.595 node1 2048kB 0 / 0 00:00:39.595 00:00:39.595 Type BDF Vendor Device NUMA Driver Device Block devices 00:00:39.595 I/OAT 0000:00:01.0 8086 0b00 0 ioatdma - - 00:00:39.595 I/OAT 0000:00:01.1 8086 0b00 0 ioatdma - - 00:00:39.595 I/OAT 0000:00:01.2 8086 0b00 0 ioatdma - - 00:00:39.595 I/OAT 0000:00:01.3 8086 0b00 0 ioatdma - - 00:00:39.595 I/OAT 0000:00:01.4 8086 0b00 0 ioatdma - - 00:00:39.595 I/OAT 0000:00:01.5 8086 0b00 0 ioatdma - - 00:00:39.595 I/OAT 0000:00:01.6 8086 0b00 0 ioatdma - - 00:00:39.595 I/OAT 0000:00:01.7 8086 0b00 0 ioatdma - - 00:00:39.595 NVMe 0000:65:00.0 144d a80a 0 nvme nvme0 nvme0n1 00:00:39.595 I/OAT 0000:80:01.0 8086 0b00 1 ioatdma - - 00:00:39.595 I/OAT 0000:80:01.1 8086 0b00 1 ioatdma - - 00:00:39.595 I/OAT 0000:80:01.2 8086 0b00 1 ioatdma - - 00:00:39.595 I/OAT 0000:80:01.3 8086 0b00 1 ioatdma - - 00:00:39.595 I/OAT 0000:80:01.4 8086 0b00 1 ioatdma - - 00:00:39.595 I/OAT 0000:80:01.5 8086 0b00 1 ioatdma - - 00:00:39.595 I/OAT 0000:80:01.6 8086 0b00 1 ioatdma - - 00:00:39.595 I/OAT 0000:80:01.7 8086 0b00 1 ioatdma - - 00:00:39.595 + rm -f /tmp/spdk-ld-path 00:00:39.595 + source autorun-spdk.conf 00:00:39.595 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:00:39.595 ++ SPDK_TEST_NVMF=1 00:00:39.595 ++ SPDK_TEST_NVME_CLI=1 00:00:39.595 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:39.595 ++ SPDK_TEST_NVMF_NICS=e810 00:00:39.595 ++ SPDK_TEST_VFIOUSER=1 00:00:39.595 ++ SPDK_RUN_UBSAN=1 00:00:39.595 ++ NET_TYPE=phy 00:00:39.595 ++ RUN_NIGHTLY=0 00:00:39.595 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:00:39.595 + [[ -n '' ]] 00:00:39.595 + sudo git config --global --add safe.directory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:39.595 + for M in /var/spdk/build-*-manifest.txt 00:00:39.595 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:00:39.595 + cp /var/spdk/build-kernel-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:00:39.595 + for M in /var/spdk/build-*-manifest.txt 00:00:39.595 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:00:39.595 + cp /var/spdk/build-pkg-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:00:39.595 + for M in /var/spdk/build-*-manifest.txt 00:00:39.595 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:00:39.595 + cp /var/spdk/build-repo-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:00:39.595 ++ uname 00:00:39.595 + [[ Linux == \L\i\n\u\x ]] 00:00:39.595 + sudo dmesg -T 00:00:39.595 + sudo dmesg --clear 00:00:39.595 + dmesg_pid=3911380 00:00:39.595 + [[ Fedora Linux == FreeBSD ]] 00:00:39.595 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:00:39.595 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:00:39.595 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:00:39.595 + [[ -x /usr/src/fio-static/fio ]] 00:00:39.595 + export FIO_BIN=/usr/src/fio-static/fio 00:00:39.595 + FIO_BIN=/usr/src/fio-static/fio 00:00:39.595 + sudo dmesg -Tw 00:00:39.595 + [[ '' == \/\v\a\r\/\j\e\n\k\i\n\s\/\w\o\r\k\s\p\a\c\e\/\n\v\m\f\-\t\c\p\-\p\h\y\-\a\u\t\o\t\e\s\t\/\q\e\m\u\_\v\f\i\o\/* ]] 00:00:39.595 + [[ ! -v VFIO_QEMU_BIN ]] 00:00:39.595 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:00:39.595 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:00:39.595 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:00:39.595 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:00:39.595 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:00:39.595 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:00:39.596 + spdk/autorun.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:00:39.596 Test configuration: 00:00:39.596 SPDK_RUN_FUNCTIONAL_TEST=1 00:00:39.596 SPDK_TEST_NVMF=1 00:00:39.596 SPDK_TEST_NVME_CLI=1 00:00:39.596 SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:39.596 SPDK_TEST_NVMF_NICS=e810 00:00:39.596 SPDK_TEST_VFIOUSER=1 00:00:39.596 SPDK_RUN_UBSAN=1 00:00:39.596 NET_TYPE=phy 00:00:39.858 RUN_NIGHTLY=0 17:21:47 -- common/autotest_common.sh@1690 -- $ [[ n == y ]] 00:00:39.858 17:21:47 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:00:39.858 17:21:47 -- scripts/common.sh@15 -- $ shopt -s extglob 00:00:39.858 17:21:47 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:00:39.858 17:21:47 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:00:39.858 17:21:47 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:00:39.858 17:21:47 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:39.858 17:21:47 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:39.858 17:21:47 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:39.858 17:21:47 -- paths/export.sh@5 -- $ export PATH 00:00:39.858 17:21:47 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:39.858 17:21:47 -- common/autobuild_common.sh@485 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:00:39.858 17:21:47 -- common/autobuild_common.sh@486 -- $ date +%s 00:00:39.858 17:21:47 -- common/autobuild_common.sh@486 -- $ mktemp -dt spdk_1729178507.XXXXXX 00:00:39.858 17:21:47 -- common/autobuild_common.sh@486 -- $ SPDK_WORKSPACE=/tmp/spdk_1729178507.6kvPVO 00:00:39.858 17:21:47 -- common/autobuild_common.sh@488 -- $ [[ -n '' ]] 00:00:39.858 17:21:47 -- common/autobuild_common.sh@492 -- $ '[' -n '' ']' 00:00:39.858 17:21:47 -- common/autobuild_common.sh@495 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:00:39.858 17:21:47 -- common/autobuild_common.sh@499 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:00:39.858 17:21:47 -- common/autobuild_common.sh@501 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:00:39.858 17:21:47 -- common/autobuild_common.sh@502 -- $ get_config_params 00:00:39.858 17:21:47 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:00:39.858 17:21:47 -- common/autotest_common.sh@10 -- $ set +x 00:00:39.858 17:21:47 -- common/autobuild_common.sh@502 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:00:39.858 17:21:47 -- common/autobuild_common.sh@504 -- $ start_monitor_resources 00:00:39.858 17:21:47 -- pm/common@17 -- $ local monitor 00:00:39.858 17:21:47 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:39.858 17:21:47 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:39.858 17:21:47 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:39.858 17:21:47 -- pm/common@21 -- $ date +%s 00:00:39.858 17:21:47 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:39.858 17:21:47 -- pm/common@25 -- $ sleep 1 00:00:39.858 17:21:47 -- pm/common@21 -- $ date +%s 00:00:39.858 17:21:47 -- pm/common@21 -- $ date +%s 00:00:39.858 17:21:47 -- pm/common@21 -- $ date +%s 00:00:39.858 17:21:47 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1729178507 00:00:39.858 17:21:47 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1729178507 00:00:39.858 17:21:47 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1729178507 00:00:39.858 17:21:47 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1729178507 00:00:39.858 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1729178507_collect-cpu-load.pm.log 00:00:39.858 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1729178507_collect-vmstat.pm.log 00:00:39.858 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1729178507_collect-cpu-temp.pm.log 00:00:39.858 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1729178507_collect-bmc-pm.bmc.pm.log 00:00:40.804 17:21:48 -- common/autobuild_common.sh@505 -- $ trap stop_monitor_resources EXIT 00:00:40.804 17:21:48 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:00:40.804 17:21:48 -- spdk/autobuild.sh@12 -- $ umask 022 00:00:40.804 17:21:48 -- spdk/autobuild.sh@13 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:40.804 17:21:48 -- spdk/autobuild.sh@16 -- $ date -u 00:00:40.804 Thu Oct 17 03:21:48 PM UTC 2024 00:00:40.804 17:21:48 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:00:40.804 v25.01-pre-91-g18d1cad7b 00:00:40.804 17:21:48 -- spdk/autobuild.sh@19 -- $ '[' 0 -eq 1 ']' 00:00:40.804 17:21:48 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:00:40.804 17:21:48 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:00:40.804 17:21:48 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:00:40.804 17:21:48 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:00:40.804 17:21:48 -- common/autotest_common.sh@10 -- $ set +x 00:00:40.804 ************************************ 00:00:40.804 START TEST ubsan 00:00:40.804 ************************************ 00:00:40.804 17:21:48 ubsan -- common/autotest_common.sh@1125 -- $ echo 'using ubsan' 00:00:40.804 using ubsan 00:00:40.804 00:00:40.804 real 0m0.001s 00:00:40.804 user 0m0.000s 00:00:40.804 sys 0m0.000s 00:00:40.804 17:21:48 ubsan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:00:40.804 17:21:48 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:00:40.804 ************************************ 00:00:40.804 END TEST ubsan 00:00:40.804 ************************************ 00:00:40.804 17:21:48 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:00:40.804 17:21:48 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:00:40.804 17:21:48 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:00:40.805 17:21:48 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:00:40.805 17:21:48 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:00:40.805 17:21:48 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:00:40.805 17:21:48 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:00:40.805 17:21:48 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:00:40.805 17:21:48 -- spdk/autobuild.sh@67 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user --with-shared 00:00:41.066 Using default SPDK env in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:00:41.066 Using default DPDK in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:00:41.328 Using 'verbs' RDMA provider 00:00:57.192 Configuring ISA-L (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal.log)...done. 00:01:09.433 Configuring ISA-L-crypto (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal-crypto.log)...done. 00:01:10.267 Creating mk/config.mk...done. 00:01:10.267 Creating mk/cc.flags.mk...done. 00:01:10.267 Type 'make' to build. 00:01:10.267 17:22:17 -- spdk/autobuild.sh@70 -- $ run_test make make -j144 00:01:10.267 17:22:17 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:01:10.267 17:22:17 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:01:10.267 17:22:18 -- common/autotest_common.sh@10 -- $ set +x 00:01:10.267 ************************************ 00:01:10.267 START TEST make 00:01:10.267 ************************************ 00:01:10.267 17:22:18 make -- common/autotest_common.sh@1125 -- $ make -j144 00:01:10.839 make[1]: Nothing to be done for 'all'. 00:01:12.228 The Meson build system 00:01:12.228 Version: 1.5.0 00:01:12.228 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user 00:01:12.228 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:12.228 Build type: native build 00:01:12.228 Project name: libvfio-user 00:01:12.228 Project version: 0.0.1 00:01:12.228 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:01:12.228 C linker for the host machine: cc ld.bfd 2.40-14 00:01:12.228 Host machine cpu family: x86_64 00:01:12.228 Host machine cpu: x86_64 00:01:12.228 Run-time dependency threads found: YES 00:01:12.228 Library dl found: YES 00:01:12.228 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:01:12.228 Run-time dependency json-c found: YES 0.17 00:01:12.228 Run-time dependency cmocka found: YES 1.1.7 00:01:12.228 Program pytest-3 found: NO 00:01:12.228 Program flake8 found: NO 00:01:12.228 Program misspell-fixer found: NO 00:01:12.228 Program restructuredtext-lint found: NO 00:01:12.228 Program valgrind found: YES (/usr/bin/valgrind) 00:01:12.228 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:01:12.228 Compiler for C supports arguments -Wmissing-declarations: YES 00:01:12.228 Compiler for C supports arguments -Wwrite-strings: YES 00:01:12.228 ../libvfio-user/test/meson.build:20: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:01:12.228 Program test-lspci.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-lspci.sh) 00:01:12.228 Program test-linkage.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-linkage.sh) 00:01:12.228 ../libvfio-user/test/py/meson.build:16: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:01:12.228 Build targets in project: 8 00:01:12.228 WARNING: Project specifies a minimum meson_version '>= 0.53.0' but uses features which were added in newer versions: 00:01:12.228 * 0.57.0: {'exclude_suites arg in add_test_setup'} 00:01:12.228 00:01:12.228 libvfio-user 0.0.1 00:01:12.228 00:01:12.228 User defined options 00:01:12.228 buildtype : debug 00:01:12.228 default_library: shared 00:01:12.228 libdir : /usr/local/lib 00:01:12.228 00:01:12.228 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:01:12.799 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:01:12.799 [1/37] Compiling C object samples/shadow_ioeventfd_server.p/shadow_ioeventfd_server.c.o 00:01:12.799 [2/37] Compiling C object samples/lspci.p/lspci.c.o 00:01:12.799 [3/37] Compiling C object samples/client.p/.._lib_tran.c.o 00:01:12.799 [4/37] Compiling C object samples/gpio-pci-idio-16.p/gpio-pci-idio-16.c.o 00:01:12.799 [5/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran.c.o 00:01:12.799 [6/37] Compiling C object samples/null.p/null.c.o 00:01:12.799 [7/37] Compiling C object test/unit_tests.p/.._lib_tran_pipe.c.o 00:01:12.799 [8/37] Compiling C object test/unit_tests.p/.._lib_tran.c.o 00:01:12.799 [9/37] Compiling C object samples/client.p/.._lib_migration.c.o 00:01:12.799 [10/37] Compiling C object lib/libvfio-user.so.0.0.1.p/irq.c.o 00:01:12.799 [11/37] Compiling C object lib/libvfio-user.so.0.0.1.p/migration.c.o 00:01:12.799 [12/37] Compiling C object test/unit_tests.p/.._lib_irq.c.o 00:01:12.799 [13/37] Compiling C object test/unit_tests.p/.._lib_pci.c.o 00:01:12.799 [14/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci.c.o 00:01:12.799 [15/37] Compiling C object test/unit_tests.p/.._lib_migration.c.o 00:01:12.799 [16/37] Compiling C object test/unit_tests.p/mocks.c.o 00:01:12.799 [17/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci_caps.c.o 00:01:12.799 [18/37] Compiling C object test/unit_tests.p/.._lib_dma.c.o 00:01:12.799 [19/37] Compiling C object samples/server.p/server.c.o 00:01:12.799 [20/37] Compiling C object samples/client.p/.._lib_tran_sock.c.o 00:01:12.799 [21/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran_sock.c.o 00:01:12.799 [22/37] Compiling C object test/unit_tests.p/.._lib_tran_sock.c.o 00:01:12.799 [23/37] Compiling C object lib/libvfio-user.so.0.0.1.p/dma.c.o 00:01:12.799 [24/37] Compiling C object test/unit_tests.p/.._lib_pci_caps.c.o 00:01:12.799 [25/37] Compiling C object test/unit_tests.p/unit-tests.c.o 00:01:12.799 [26/37] Compiling C object samples/client.p/client.c.o 00:01:12.799 [27/37] Linking target samples/client 00:01:12.799 [28/37] Compiling C object lib/libvfio-user.so.0.0.1.p/libvfio-user.c.o 00:01:12.799 [29/37] Compiling C object test/unit_tests.p/.._lib_libvfio-user.c.o 00:01:13.060 [30/37] Linking target lib/libvfio-user.so.0.0.1 00:01:13.060 [31/37] Linking target test/unit_tests 00:01:13.060 [32/37] Generating symbol file lib/libvfio-user.so.0.0.1.p/libvfio-user.so.0.0.1.symbols 00:01:13.060 [33/37] Linking target samples/server 00:01:13.060 [34/37] Linking target samples/gpio-pci-idio-16 00:01:13.060 [35/37] Linking target samples/lspci 00:01:13.060 [36/37] Linking target samples/shadow_ioeventfd_server 00:01:13.060 [37/37] Linking target samples/null 00:01:13.060 INFO: autodetecting backend as ninja 00:01:13.060 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:13.322 DESTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user meson install --quiet -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:13.583 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:01:13.583 ninja: no work to do. 00:01:20.174 The Meson build system 00:01:20.174 Version: 1.5.0 00:01:20.174 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk 00:01:20.174 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp 00:01:20.174 Build type: native build 00:01:20.174 Program cat found: YES (/usr/bin/cat) 00:01:20.174 Project name: DPDK 00:01:20.174 Project version: 24.03.0 00:01:20.174 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:01:20.174 C linker for the host machine: cc ld.bfd 2.40-14 00:01:20.174 Host machine cpu family: x86_64 00:01:20.174 Host machine cpu: x86_64 00:01:20.174 Message: ## Building in Developer Mode ## 00:01:20.174 Program pkg-config found: YES (/usr/bin/pkg-config) 00:01:20.174 Program check-symbols.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/check-symbols.sh) 00:01:20.174 Program options-ibverbs-static.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:01:20.174 Program python3 found: YES (/usr/bin/python3) 00:01:20.174 Program cat found: YES (/usr/bin/cat) 00:01:20.174 Compiler for C supports arguments -march=native: YES 00:01:20.174 Checking for size of "void *" : 8 00:01:20.174 Checking for size of "void *" : 8 (cached) 00:01:20.174 Compiler for C supports link arguments -Wl,--undefined-version: YES 00:01:20.174 Library m found: YES 00:01:20.174 Library numa found: YES 00:01:20.174 Has header "numaif.h" : YES 00:01:20.174 Library fdt found: NO 00:01:20.174 Library execinfo found: NO 00:01:20.174 Has header "execinfo.h" : YES 00:01:20.174 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:01:20.174 Run-time dependency libarchive found: NO (tried pkgconfig) 00:01:20.174 Run-time dependency libbsd found: NO (tried pkgconfig) 00:01:20.174 Run-time dependency jansson found: NO (tried pkgconfig) 00:01:20.174 Run-time dependency openssl found: YES 3.1.1 00:01:20.174 Run-time dependency libpcap found: YES 1.10.4 00:01:20.174 Has header "pcap.h" with dependency libpcap: YES 00:01:20.174 Compiler for C supports arguments -Wcast-qual: YES 00:01:20.174 Compiler for C supports arguments -Wdeprecated: YES 00:01:20.174 Compiler for C supports arguments -Wformat: YES 00:01:20.174 Compiler for C supports arguments -Wformat-nonliteral: NO 00:01:20.174 Compiler for C supports arguments -Wformat-security: NO 00:01:20.174 Compiler for C supports arguments -Wmissing-declarations: YES 00:01:20.174 Compiler for C supports arguments -Wmissing-prototypes: YES 00:01:20.174 Compiler for C supports arguments -Wnested-externs: YES 00:01:20.174 Compiler for C supports arguments -Wold-style-definition: YES 00:01:20.174 Compiler for C supports arguments -Wpointer-arith: YES 00:01:20.174 Compiler for C supports arguments -Wsign-compare: YES 00:01:20.174 Compiler for C supports arguments -Wstrict-prototypes: YES 00:01:20.174 Compiler for C supports arguments -Wundef: YES 00:01:20.174 Compiler for C supports arguments -Wwrite-strings: YES 00:01:20.174 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:01:20.174 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:01:20.174 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:01:20.174 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:01:20.174 Program objdump found: YES (/usr/bin/objdump) 00:01:20.174 Compiler for C supports arguments -mavx512f: YES 00:01:20.174 Checking if "AVX512 checking" compiles: YES 00:01:20.174 Fetching value of define "__SSE4_2__" : 1 00:01:20.174 Fetching value of define "__AES__" : 1 00:01:20.174 Fetching value of define "__AVX__" : 1 00:01:20.174 Fetching value of define "__AVX2__" : 1 00:01:20.174 Fetching value of define "__AVX512BW__" : 1 00:01:20.174 Fetching value of define "__AVX512CD__" : 1 00:01:20.174 Fetching value of define "__AVX512DQ__" : 1 00:01:20.174 Fetching value of define "__AVX512F__" : 1 00:01:20.174 Fetching value of define "__AVX512VL__" : 1 00:01:20.174 Fetching value of define "__PCLMUL__" : 1 00:01:20.175 Fetching value of define "__RDRND__" : 1 00:01:20.175 Fetching value of define "__RDSEED__" : 1 00:01:20.175 Fetching value of define "__VPCLMULQDQ__" : 1 00:01:20.175 Fetching value of define "__znver1__" : (undefined) 00:01:20.175 Fetching value of define "__znver2__" : (undefined) 00:01:20.175 Fetching value of define "__znver3__" : (undefined) 00:01:20.175 Fetching value of define "__znver4__" : (undefined) 00:01:20.175 Compiler for C supports arguments -Wno-format-truncation: YES 00:01:20.175 Message: lib/log: Defining dependency "log" 00:01:20.175 Message: lib/kvargs: Defining dependency "kvargs" 00:01:20.175 Message: lib/telemetry: Defining dependency "telemetry" 00:01:20.175 Checking for function "getentropy" : NO 00:01:20.175 Message: lib/eal: Defining dependency "eal" 00:01:20.175 Message: lib/ring: Defining dependency "ring" 00:01:20.175 Message: lib/rcu: Defining dependency "rcu" 00:01:20.175 Message: lib/mempool: Defining dependency "mempool" 00:01:20.175 Message: lib/mbuf: Defining dependency "mbuf" 00:01:20.175 Fetching value of define "__PCLMUL__" : 1 (cached) 00:01:20.175 Fetching value of define "__AVX512F__" : 1 (cached) 00:01:20.175 Fetching value of define "__AVX512BW__" : 1 (cached) 00:01:20.175 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:01:20.175 Fetching value of define "__AVX512VL__" : 1 (cached) 00:01:20.175 Fetching value of define "__VPCLMULQDQ__" : 1 (cached) 00:01:20.175 Compiler for C supports arguments -mpclmul: YES 00:01:20.175 Compiler for C supports arguments -maes: YES 00:01:20.175 Compiler for C supports arguments -mavx512f: YES (cached) 00:01:20.175 Compiler for C supports arguments -mavx512bw: YES 00:01:20.175 Compiler for C supports arguments -mavx512dq: YES 00:01:20.175 Compiler for C supports arguments -mavx512vl: YES 00:01:20.175 Compiler for C supports arguments -mvpclmulqdq: YES 00:01:20.175 Compiler for C supports arguments -mavx2: YES 00:01:20.175 Compiler for C supports arguments -mavx: YES 00:01:20.175 Message: lib/net: Defining dependency "net" 00:01:20.175 Message: lib/meter: Defining dependency "meter" 00:01:20.175 Message: lib/ethdev: Defining dependency "ethdev" 00:01:20.175 Message: lib/pci: Defining dependency "pci" 00:01:20.175 Message: lib/cmdline: Defining dependency "cmdline" 00:01:20.175 Message: lib/hash: Defining dependency "hash" 00:01:20.175 Message: lib/timer: Defining dependency "timer" 00:01:20.175 Message: lib/compressdev: Defining dependency "compressdev" 00:01:20.175 Message: lib/cryptodev: Defining dependency "cryptodev" 00:01:20.175 Message: lib/dmadev: Defining dependency "dmadev" 00:01:20.175 Compiler for C supports arguments -Wno-cast-qual: YES 00:01:20.175 Message: lib/power: Defining dependency "power" 00:01:20.175 Message: lib/reorder: Defining dependency "reorder" 00:01:20.175 Message: lib/security: Defining dependency "security" 00:01:20.175 Has header "linux/userfaultfd.h" : YES 00:01:20.175 Has header "linux/vduse.h" : YES 00:01:20.175 Message: lib/vhost: Defining dependency "vhost" 00:01:20.175 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:01:20.175 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:01:20.175 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:01:20.175 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:01:20.175 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:01:20.175 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:01:20.175 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:01:20.175 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:01:20.175 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:01:20.175 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:01:20.175 Program doxygen found: YES (/usr/local/bin/doxygen) 00:01:20.175 Configuring doxy-api-html.conf using configuration 00:01:20.175 Configuring doxy-api-man.conf using configuration 00:01:20.175 Program mandb found: YES (/usr/bin/mandb) 00:01:20.175 Program sphinx-build found: NO 00:01:20.175 Configuring rte_build_config.h using configuration 00:01:20.175 Message: 00:01:20.175 ================= 00:01:20.175 Applications Enabled 00:01:20.175 ================= 00:01:20.175 00:01:20.175 apps: 00:01:20.175 00:01:20.175 00:01:20.175 Message: 00:01:20.175 ================= 00:01:20.175 Libraries Enabled 00:01:20.175 ================= 00:01:20.175 00:01:20.175 libs: 00:01:20.175 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:01:20.175 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:01:20.175 cryptodev, dmadev, power, reorder, security, vhost, 00:01:20.175 00:01:20.175 Message: 00:01:20.175 =============== 00:01:20.175 Drivers Enabled 00:01:20.175 =============== 00:01:20.175 00:01:20.175 common: 00:01:20.175 00:01:20.175 bus: 00:01:20.175 pci, vdev, 00:01:20.175 mempool: 00:01:20.175 ring, 00:01:20.175 dma: 00:01:20.175 00:01:20.175 net: 00:01:20.175 00:01:20.175 crypto: 00:01:20.175 00:01:20.175 compress: 00:01:20.175 00:01:20.175 vdpa: 00:01:20.175 00:01:20.175 00:01:20.175 Message: 00:01:20.175 ================= 00:01:20.175 Content Skipped 00:01:20.175 ================= 00:01:20.175 00:01:20.175 apps: 00:01:20.175 dumpcap: explicitly disabled via build config 00:01:20.175 graph: explicitly disabled via build config 00:01:20.175 pdump: explicitly disabled via build config 00:01:20.175 proc-info: explicitly disabled via build config 00:01:20.175 test-acl: explicitly disabled via build config 00:01:20.175 test-bbdev: explicitly disabled via build config 00:01:20.175 test-cmdline: explicitly disabled via build config 00:01:20.175 test-compress-perf: explicitly disabled via build config 00:01:20.175 test-crypto-perf: explicitly disabled via build config 00:01:20.175 test-dma-perf: explicitly disabled via build config 00:01:20.175 test-eventdev: explicitly disabled via build config 00:01:20.175 test-fib: explicitly disabled via build config 00:01:20.175 test-flow-perf: explicitly disabled via build config 00:01:20.175 test-gpudev: explicitly disabled via build config 00:01:20.175 test-mldev: explicitly disabled via build config 00:01:20.175 test-pipeline: explicitly disabled via build config 00:01:20.175 test-pmd: explicitly disabled via build config 00:01:20.175 test-regex: explicitly disabled via build config 00:01:20.175 test-sad: explicitly disabled via build config 00:01:20.175 test-security-perf: explicitly disabled via build config 00:01:20.175 00:01:20.175 libs: 00:01:20.175 argparse: explicitly disabled via build config 00:01:20.175 metrics: explicitly disabled via build config 00:01:20.175 acl: explicitly disabled via build config 00:01:20.175 bbdev: explicitly disabled via build config 00:01:20.175 bitratestats: explicitly disabled via build config 00:01:20.175 bpf: explicitly disabled via build config 00:01:20.175 cfgfile: explicitly disabled via build config 00:01:20.175 distributor: explicitly disabled via build config 00:01:20.175 efd: explicitly disabled via build config 00:01:20.175 eventdev: explicitly disabled via build config 00:01:20.175 dispatcher: explicitly disabled via build config 00:01:20.175 gpudev: explicitly disabled via build config 00:01:20.175 gro: explicitly disabled via build config 00:01:20.175 gso: explicitly disabled via build config 00:01:20.175 ip_frag: explicitly disabled via build config 00:01:20.175 jobstats: explicitly disabled via build config 00:01:20.175 latencystats: explicitly disabled via build config 00:01:20.175 lpm: explicitly disabled via build config 00:01:20.175 member: explicitly disabled via build config 00:01:20.175 pcapng: explicitly disabled via build config 00:01:20.175 rawdev: explicitly disabled via build config 00:01:20.175 regexdev: explicitly disabled via build config 00:01:20.175 mldev: explicitly disabled via build config 00:01:20.175 rib: explicitly disabled via build config 00:01:20.175 sched: explicitly disabled via build config 00:01:20.175 stack: explicitly disabled via build config 00:01:20.175 ipsec: explicitly disabled via build config 00:01:20.175 pdcp: explicitly disabled via build config 00:01:20.175 fib: explicitly disabled via build config 00:01:20.175 port: explicitly disabled via build config 00:01:20.175 pdump: explicitly disabled via build config 00:01:20.175 table: explicitly disabled via build config 00:01:20.175 pipeline: explicitly disabled via build config 00:01:20.175 graph: explicitly disabled via build config 00:01:20.175 node: explicitly disabled via build config 00:01:20.175 00:01:20.175 drivers: 00:01:20.175 common/cpt: not in enabled drivers build config 00:01:20.175 common/dpaax: not in enabled drivers build config 00:01:20.175 common/iavf: not in enabled drivers build config 00:01:20.175 common/idpf: not in enabled drivers build config 00:01:20.175 common/ionic: not in enabled drivers build config 00:01:20.175 common/mvep: not in enabled drivers build config 00:01:20.175 common/octeontx: not in enabled drivers build config 00:01:20.175 bus/auxiliary: not in enabled drivers build config 00:01:20.175 bus/cdx: not in enabled drivers build config 00:01:20.175 bus/dpaa: not in enabled drivers build config 00:01:20.175 bus/fslmc: not in enabled drivers build config 00:01:20.175 bus/ifpga: not in enabled drivers build config 00:01:20.175 bus/platform: not in enabled drivers build config 00:01:20.175 bus/uacce: not in enabled drivers build config 00:01:20.175 bus/vmbus: not in enabled drivers build config 00:01:20.175 common/cnxk: not in enabled drivers build config 00:01:20.175 common/mlx5: not in enabled drivers build config 00:01:20.175 common/nfp: not in enabled drivers build config 00:01:20.175 common/nitrox: not in enabled drivers build config 00:01:20.175 common/qat: not in enabled drivers build config 00:01:20.175 common/sfc_efx: not in enabled drivers build config 00:01:20.175 mempool/bucket: not in enabled drivers build config 00:01:20.175 mempool/cnxk: not in enabled drivers build config 00:01:20.175 mempool/dpaa: not in enabled drivers build config 00:01:20.175 mempool/dpaa2: not in enabled drivers build config 00:01:20.175 mempool/octeontx: not in enabled drivers build config 00:01:20.175 mempool/stack: not in enabled drivers build config 00:01:20.175 dma/cnxk: not in enabled drivers build config 00:01:20.175 dma/dpaa: not in enabled drivers build config 00:01:20.175 dma/dpaa2: not in enabled drivers build config 00:01:20.175 dma/hisilicon: not in enabled drivers build config 00:01:20.175 dma/idxd: not in enabled drivers build config 00:01:20.175 dma/ioat: not in enabled drivers build config 00:01:20.175 dma/skeleton: not in enabled drivers build config 00:01:20.175 net/af_packet: not in enabled drivers build config 00:01:20.175 net/af_xdp: not in enabled drivers build config 00:01:20.175 net/ark: not in enabled drivers build config 00:01:20.175 net/atlantic: not in enabled drivers build config 00:01:20.175 net/avp: not in enabled drivers build config 00:01:20.175 net/axgbe: not in enabled drivers build config 00:01:20.175 net/bnx2x: not in enabled drivers build config 00:01:20.175 net/bnxt: not in enabled drivers build config 00:01:20.175 net/bonding: not in enabled drivers build config 00:01:20.175 net/cnxk: not in enabled drivers build config 00:01:20.176 net/cpfl: not in enabled drivers build config 00:01:20.176 net/cxgbe: not in enabled drivers build config 00:01:20.176 net/dpaa: not in enabled drivers build config 00:01:20.176 net/dpaa2: not in enabled drivers build config 00:01:20.176 net/e1000: not in enabled drivers build config 00:01:20.176 net/ena: not in enabled drivers build config 00:01:20.176 net/enetc: not in enabled drivers build config 00:01:20.176 net/enetfec: not in enabled drivers build config 00:01:20.176 net/enic: not in enabled drivers build config 00:01:20.176 net/failsafe: not in enabled drivers build config 00:01:20.176 net/fm10k: not in enabled drivers build config 00:01:20.176 net/gve: not in enabled drivers build config 00:01:20.176 net/hinic: not in enabled drivers build config 00:01:20.176 net/hns3: not in enabled drivers build config 00:01:20.176 net/i40e: not in enabled drivers build config 00:01:20.176 net/iavf: not in enabled drivers build config 00:01:20.176 net/ice: not in enabled drivers build config 00:01:20.176 net/idpf: not in enabled drivers build config 00:01:20.176 net/igc: not in enabled drivers build config 00:01:20.176 net/ionic: not in enabled drivers build config 00:01:20.176 net/ipn3ke: not in enabled drivers build config 00:01:20.176 net/ixgbe: not in enabled drivers build config 00:01:20.176 net/mana: not in enabled drivers build config 00:01:20.176 net/memif: not in enabled drivers build config 00:01:20.176 net/mlx4: not in enabled drivers build config 00:01:20.176 net/mlx5: not in enabled drivers build config 00:01:20.176 net/mvneta: not in enabled drivers build config 00:01:20.176 net/mvpp2: not in enabled drivers build config 00:01:20.176 net/netvsc: not in enabled drivers build config 00:01:20.176 net/nfb: not in enabled drivers build config 00:01:20.176 net/nfp: not in enabled drivers build config 00:01:20.176 net/ngbe: not in enabled drivers build config 00:01:20.176 net/null: not in enabled drivers build config 00:01:20.176 net/octeontx: not in enabled drivers build config 00:01:20.176 net/octeon_ep: not in enabled drivers build config 00:01:20.176 net/pcap: not in enabled drivers build config 00:01:20.176 net/pfe: not in enabled drivers build config 00:01:20.176 net/qede: not in enabled drivers build config 00:01:20.176 net/ring: not in enabled drivers build config 00:01:20.176 net/sfc: not in enabled drivers build config 00:01:20.176 net/softnic: not in enabled drivers build config 00:01:20.176 net/tap: not in enabled drivers build config 00:01:20.176 net/thunderx: not in enabled drivers build config 00:01:20.176 net/txgbe: not in enabled drivers build config 00:01:20.176 net/vdev_netvsc: not in enabled drivers build config 00:01:20.176 net/vhost: not in enabled drivers build config 00:01:20.176 net/virtio: not in enabled drivers build config 00:01:20.176 net/vmxnet3: not in enabled drivers build config 00:01:20.176 raw/*: missing internal dependency, "rawdev" 00:01:20.176 crypto/armv8: not in enabled drivers build config 00:01:20.176 crypto/bcmfs: not in enabled drivers build config 00:01:20.176 crypto/caam_jr: not in enabled drivers build config 00:01:20.176 crypto/ccp: not in enabled drivers build config 00:01:20.176 crypto/cnxk: not in enabled drivers build config 00:01:20.176 crypto/dpaa_sec: not in enabled drivers build config 00:01:20.176 crypto/dpaa2_sec: not in enabled drivers build config 00:01:20.176 crypto/ipsec_mb: not in enabled drivers build config 00:01:20.176 crypto/mlx5: not in enabled drivers build config 00:01:20.176 crypto/mvsam: not in enabled drivers build config 00:01:20.176 crypto/nitrox: not in enabled drivers build config 00:01:20.176 crypto/null: not in enabled drivers build config 00:01:20.176 crypto/octeontx: not in enabled drivers build config 00:01:20.176 crypto/openssl: not in enabled drivers build config 00:01:20.176 crypto/scheduler: not in enabled drivers build config 00:01:20.176 crypto/uadk: not in enabled drivers build config 00:01:20.176 crypto/virtio: not in enabled drivers build config 00:01:20.176 compress/isal: not in enabled drivers build config 00:01:20.176 compress/mlx5: not in enabled drivers build config 00:01:20.176 compress/nitrox: not in enabled drivers build config 00:01:20.176 compress/octeontx: not in enabled drivers build config 00:01:20.176 compress/zlib: not in enabled drivers build config 00:01:20.176 regex/*: missing internal dependency, "regexdev" 00:01:20.176 ml/*: missing internal dependency, "mldev" 00:01:20.176 vdpa/ifc: not in enabled drivers build config 00:01:20.176 vdpa/mlx5: not in enabled drivers build config 00:01:20.176 vdpa/nfp: not in enabled drivers build config 00:01:20.176 vdpa/sfc: not in enabled drivers build config 00:01:20.176 event/*: missing internal dependency, "eventdev" 00:01:20.176 baseband/*: missing internal dependency, "bbdev" 00:01:20.176 gpu/*: missing internal dependency, "gpudev" 00:01:20.176 00:01:20.176 00:01:20.176 Build targets in project: 84 00:01:20.176 00:01:20.176 DPDK 24.03.0 00:01:20.176 00:01:20.176 User defined options 00:01:20.176 buildtype : debug 00:01:20.176 default_library : shared 00:01:20.176 libdir : lib 00:01:20.176 prefix : /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:01:20.176 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:01:20.176 c_link_args : 00:01:20.176 cpu_instruction_set: native 00:01:20.176 disable_apps : test-fib,test-sad,test,test-regex,test-security-perf,test-bbdev,dumpcap,test-crypto-perf,test-flow-perf,test-gpudev,test-cmdline,test-dma-perf,test-eventdev,test-pipeline,test-acl,proc-info,test-compress-perf,graph,test-pmd,test-mldev,pdump 00:01:20.176 disable_libs : bbdev,argparse,latencystats,member,gpudev,mldev,pipeline,lpm,efd,regexdev,sched,node,dispatcher,table,bpf,port,gro,fib,cfgfile,ip_frag,gso,rawdev,ipsec,pdcp,rib,acl,metrics,graph,pcapng,jobstats,eventdev,stack,bitratestats,distributor,pdump 00:01:20.176 enable_docs : false 00:01:20.176 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring 00:01:20.176 enable_kmods : false 00:01:20.176 max_lcores : 128 00:01:20.176 tests : false 00:01:20.176 00:01:20.176 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:01:20.176 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp' 00:01:20.176 [1/267] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:01:20.176 [2/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:01:20.176 [3/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:01:20.176 [4/267] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:01:20.176 [5/267] Linking static target lib/librte_kvargs.a 00:01:20.176 [6/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:01:20.176 [7/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:01:20.176 [8/267] Compiling C object lib/librte_log.a.p/log_log.c.o 00:01:20.176 [9/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:01:20.176 [10/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:01:20.176 [11/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:01:20.176 [12/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:01:20.176 [13/267] Linking static target lib/librte_log.a 00:01:20.176 [14/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:01:20.176 [15/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:01:20.176 [16/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:01:20.176 [17/267] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:01:20.176 [18/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:01:20.176 [19/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:01:20.176 [20/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:01:20.176 [21/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:01:20.176 [22/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:01:20.176 [23/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:01:20.176 [24/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:01:20.176 [25/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:01:20.176 [26/267] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:01:20.176 [27/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:01:20.435 [28/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:01:20.435 [29/267] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:01:20.435 [30/267] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:01:20.435 [31/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:01:20.435 [32/267] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:01:20.435 [33/267] Linking static target lib/librte_pci.a 00:01:20.435 [34/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:01:20.435 [35/267] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:01:20.435 [36/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:01:20.435 [37/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:01:20.435 [38/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:01:20.435 [39/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:01:20.435 [40/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:01:20.435 [41/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:01:20.435 [42/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:01:20.435 [43/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:01:20.435 [44/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:01:20.435 [45/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:01:20.435 [46/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:01:20.435 [47/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:01:20.695 [48/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:01:20.695 [49/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:01:20.695 [50/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:01:20.695 [51/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:01:20.695 [52/267] Compiling C object lib/librte_net.a.p/net_net_crc_avx512.c.o 00:01:20.695 [53/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:01:20.695 [54/267] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:01:20.695 [55/267] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:01:20.695 [56/267] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:01:20.695 [57/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:01:20.695 [58/267] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:01:20.695 [59/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:01:20.695 [60/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:01:20.695 [61/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:01:20.695 [62/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:01:20.695 [63/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:01:20.695 [64/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:01:20.695 [65/267] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:01:20.695 [66/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:01:20.695 [67/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:01:20.695 [68/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:01:20.695 [69/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:01:20.695 [70/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:01:20.695 [71/267] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:01:20.695 [72/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:01:20.695 [73/267] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:01:20.695 [74/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:01:20.695 [75/267] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:01:20.695 [76/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:01:20.695 [77/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:01:20.695 [78/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:01:20.695 [79/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:01:20.695 [80/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:01:20.695 [81/267] Linking static target lib/librte_cmdline.a 00:01:20.695 [82/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:01:20.695 [83/267] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:01:20.695 [84/267] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:01:20.695 [85/267] Linking static target lib/librte_timer.a 00:01:20.695 [86/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:01:20.695 [87/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:01:20.695 [88/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:01:20.695 [89/267] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:01:20.695 [90/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:01:20.695 [91/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:01:20.695 [92/267] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:01:20.695 [93/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:01:20.695 [94/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:01:20.695 [95/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:01:20.695 [96/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:01:20.695 [97/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:01:20.695 [98/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:01:20.695 [99/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:01:20.695 [100/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:01:20.695 [101/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:01:20.695 [102/267] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:01:20.695 [103/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:01:20.695 [104/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:01:20.695 [105/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:01:20.695 [106/267] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:01:20.695 [107/267] Linking static target lib/librte_dmadev.a 00:01:20.695 [108/267] Linking static target lib/librte_rcu.a 00:01:20.695 [109/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:01:20.695 [110/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:01:20.695 [111/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:01:20.695 [112/267] Linking static target drivers/libtmp_rte_bus_vdev.a 00:01:20.695 [113/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:01:20.695 [114/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:01:20.695 [115/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:01:20.695 [116/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:01:20.695 [117/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:01:20.695 [118/267] Linking static target lib/librte_telemetry.a 00:01:20.695 [119/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:01:20.695 [120/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:01:20.695 [121/267] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:01:20.695 [122/267] Linking static target lib/librte_meter.a 00:01:20.696 [123/267] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:01:20.696 [124/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:01:20.696 [125/267] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:01:20.696 [126/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:01:20.696 [127/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:01:20.696 [128/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:01:20.696 [129/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:01:20.696 [130/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:01:20.696 [131/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:01:20.696 [132/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:01:20.696 [133/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:01:20.696 [134/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:01:20.696 [135/267] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:01:20.696 [136/267] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:01:20.953 [137/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:01:20.953 [138/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:01:20.953 [139/267] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:01:20.953 [140/267] Linking static target lib/librte_mempool.a 00:01:20.953 [141/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:01:20.953 [142/267] Linking static target lib/librte_compressdev.a 00:01:20.953 [143/267] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:01:20.953 [144/267] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:01:20.953 [145/267] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:01:20.953 [146/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:01:20.953 [147/267] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:01:20.953 [148/267] Linking static target lib/librte_ring.a 00:01:20.953 [149/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:01:20.953 [150/267] Linking static target lib/librte_power.a 00:01:20.953 [151/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:01:20.953 [152/267] Linking target lib/librte_log.so.24.1 00:01:20.953 [153/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:01:20.953 [154/267] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:01:20.953 [155/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:01:20.953 [156/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:01:20.953 [157/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:01:20.953 [158/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:01:20.953 [159/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:01:20.953 [160/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:01:20.953 [161/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:01:20.953 [162/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:01:20.953 [163/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:01:20.953 [164/267] Linking static target drivers/libtmp_rte_bus_pci.a 00:01:20.953 [165/267] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:01:20.953 [166/267] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:01:20.953 [167/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:01:20.953 [168/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:01:20.953 [169/267] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:01:20.953 [170/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:01:20.953 [171/267] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:01:20.953 [172/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:01:20.953 [173/267] Linking static target drivers/librte_bus_vdev.a 00:01:20.953 [174/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:01:20.953 [175/267] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:01:20.953 [176/267] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:01:20.953 [177/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:01:20.953 [178/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:01:20.953 [179/267] Linking static target lib/librte_eal.a 00:01:20.953 [180/267] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:01:20.953 [181/267] Linking static target lib/librte_net.a 00:01:20.953 [182/267] Linking static target lib/librte_mbuf.a 00:01:20.953 [183/267] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:01:20.953 [184/267] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:01:20.953 [185/267] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:01:20.953 [186/267] Linking static target lib/librte_reorder.a 00:01:20.953 [187/267] Linking static target lib/librte_security.a 00:01:20.953 [188/267] Linking static target drivers/libtmp_rte_mempool_ring.a 00:01:20.953 [189/267] Linking target lib/librte_kvargs.so.24.1 00:01:21.213 [190/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:01:21.213 [191/267] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.213 [192/267] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:01:21.213 [193/267] Linking static target lib/librte_hash.a 00:01:21.213 [194/267] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:01:21.213 [195/267] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:01:21.213 [196/267] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.213 [197/267] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:01:21.213 [198/267] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.213 [199/267] Linking static target drivers/librte_bus_pci.a 00:01:21.213 [200/267] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:01:21.213 [201/267] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:01:21.213 [202/267] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:01:21.213 [203/267] Linking static target drivers/librte_mempool_ring.a 00:01:21.213 [204/267] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:01:21.213 [205/267] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.213 [206/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:01:21.213 [207/267] Linking static target lib/librte_cryptodev.a 00:01:21.474 [208/267] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.474 [209/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:01:21.474 [210/267] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.474 [211/267] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.474 [212/267] Linking target lib/librte_telemetry.so.24.1 00:01:21.474 [213/267] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.474 [214/267] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.474 [215/267] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:01:21.736 [216/267] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.736 [217/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:01:21.736 [218/267] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.736 [219/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:01:21.736 [220/267] Linking static target lib/librte_ethdev.a 00:01:21.998 [221/267] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.998 [222/267] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.998 [223/267] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.998 [224/267] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:01:21.998 [225/267] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:01:22.259 [226/267] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:01:22.831 [227/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:01:23.092 [228/267] Linking static target lib/librte_vhost.a 00:01:23.353 [229/267] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:25.268 [230/267] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:01:31.854 [231/267] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:32.425 [232/267] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:01:32.684 [233/267] Linking target lib/librte_eal.so.24.1 00:01:32.684 [234/267] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:01:32.684 [235/267] Linking target lib/librte_meter.so.24.1 00:01:32.684 [236/267] Linking target lib/librte_ring.so.24.1 00:01:32.684 [237/267] Linking target lib/librte_timer.so.24.1 00:01:32.684 [238/267] Linking target lib/librte_pci.so.24.1 00:01:32.684 [239/267] Linking target drivers/librte_bus_vdev.so.24.1 00:01:32.684 [240/267] Linking target lib/librte_dmadev.so.24.1 00:01:32.945 [241/267] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:01:32.945 [242/267] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:01:32.945 [243/267] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:01:32.945 [244/267] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:01:32.945 [245/267] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:01:32.945 [246/267] Linking target lib/librte_rcu.so.24.1 00:01:32.945 [247/267] Linking target lib/librte_mempool.so.24.1 00:01:32.945 [248/267] Linking target drivers/librte_bus_pci.so.24.1 00:01:33.205 [249/267] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:01:33.205 [250/267] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:01:33.205 [251/267] Linking target lib/librte_mbuf.so.24.1 00:01:33.205 [252/267] Linking target drivers/librte_mempool_ring.so.24.1 00:01:33.205 [253/267] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:01:33.465 [254/267] Linking target lib/librte_compressdev.so.24.1 00:01:33.465 [255/267] Linking target lib/librte_net.so.24.1 00:01:33.465 [256/267] Linking target lib/librte_reorder.so.24.1 00:01:33.465 [257/267] Linking target lib/librte_cryptodev.so.24.1 00:01:33.465 [258/267] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:01:33.465 [259/267] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:01:33.465 [260/267] Linking target lib/librte_hash.so.24.1 00:01:33.465 [261/267] Linking target lib/librte_cmdline.so.24.1 00:01:33.465 [262/267] Linking target lib/librte_security.so.24.1 00:01:33.465 [263/267] Linking target lib/librte_ethdev.so.24.1 00:01:33.725 [264/267] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:01:33.725 [265/267] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:01:33.725 [266/267] Linking target lib/librte_power.so.24.1 00:01:33.725 [267/267] Linking target lib/librte_vhost.so.24.1 00:01:33.725 INFO: autodetecting backend as ninja 00:01:33.725 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp -j 144 00:01:37.931 CC lib/log/log.o 00:01:37.931 CC lib/ut_mock/mock.o 00:01:37.931 CC lib/log/log_flags.o 00:01:37.931 CC lib/log/log_deprecated.o 00:01:37.931 CC lib/ut/ut.o 00:01:37.931 LIB libspdk_ut_mock.a 00:01:37.931 LIB libspdk_ut.a 00:01:37.931 LIB libspdk_log.a 00:01:37.931 SO libspdk_ut_mock.so.6.0 00:01:37.931 SO libspdk_log.so.7.1 00:01:37.931 SO libspdk_ut.so.2.0 00:01:37.931 SYMLINK libspdk_ut_mock.so 00:01:37.931 SYMLINK libspdk_log.so 00:01:37.931 SYMLINK libspdk_ut.so 00:01:37.931 CC lib/ioat/ioat.o 00:01:38.193 CC lib/util/base64.o 00:01:38.193 CC lib/util/bit_array.o 00:01:38.193 CC lib/util/crc16.o 00:01:38.193 CC lib/util/cpuset.o 00:01:38.193 CC lib/util/crc32.o 00:01:38.193 CC lib/util/crc32c.o 00:01:38.193 CC lib/util/crc32_ieee.o 00:01:38.193 CC lib/util/crc64.o 00:01:38.193 CXX lib/trace_parser/trace.o 00:01:38.193 CC lib/util/dif.o 00:01:38.193 CC lib/util/fd.o 00:01:38.193 CC lib/util/fd_group.o 00:01:38.193 CC lib/dma/dma.o 00:01:38.193 CC lib/util/file.o 00:01:38.193 CC lib/util/hexlify.o 00:01:38.193 CC lib/util/iov.o 00:01:38.193 CC lib/util/math.o 00:01:38.193 CC lib/util/net.o 00:01:38.193 CC lib/util/pipe.o 00:01:38.193 CC lib/util/strerror_tls.o 00:01:38.193 CC lib/util/string.o 00:01:38.193 CC lib/util/uuid.o 00:01:38.193 CC lib/util/xor.o 00:01:38.193 CC lib/util/zipf.o 00:01:38.193 CC lib/util/md5.o 00:01:38.193 CC lib/vfio_user/host/vfio_user_pci.o 00:01:38.193 CC lib/vfio_user/host/vfio_user.o 00:01:38.193 LIB libspdk_dma.a 00:01:38.454 SO libspdk_dma.so.5.0 00:01:38.454 LIB libspdk_ioat.a 00:01:38.454 SYMLINK libspdk_dma.so 00:01:38.454 SO libspdk_ioat.so.7.0 00:01:38.454 SYMLINK libspdk_ioat.so 00:01:38.454 LIB libspdk_vfio_user.a 00:01:38.454 SO libspdk_vfio_user.so.5.0 00:01:38.454 LIB libspdk_util.a 00:01:38.715 SYMLINK libspdk_vfio_user.so 00:01:38.715 SO libspdk_util.so.10.0 00:01:38.715 SYMLINK libspdk_util.so 00:01:38.976 LIB libspdk_trace_parser.a 00:01:38.976 SO libspdk_trace_parser.so.6.0 00:01:38.976 SYMLINK libspdk_trace_parser.so 00:01:39.236 CC lib/vmd/vmd.o 00:01:39.236 CC lib/vmd/led.o 00:01:39.236 CC lib/conf/conf.o 00:01:39.236 CC lib/json/json_parse.o 00:01:39.236 CC lib/env_dpdk/env.o 00:01:39.236 CC lib/json/json_util.o 00:01:39.236 CC lib/env_dpdk/memory.o 00:01:39.236 CC lib/rdma_utils/rdma_utils.o 00:01:39.236 CC lib/env_dpdk/pci.o 00:01:39.236 CC lib/json/json_write.o 00:01:39.236 CC lib/idxd/idxd.o 00:01:39.236 CC lib/env_dpdk/init.o 00:01:39.236 CC lib/idxd/idxd_user.o 00:01:39.236 CC lib/env_dpdk/threads.o 00:01:39.236 CC lib/idxd/idxd_kernel.o 00:01:39.236 CC lib/env_dpdk/pci_ioat.o 00:01:39.236 CC lib/env_dpdk/pci_virtio.o 00:01:39.236 CC lib/env_dpdk/pci_vmd.o 00:01:39.236 CC lib/env_dpdk/pci_idxd.o 00:01:39.236 CC lib/env_dpdk/sigbus_handler.o 00:01:39.236 CC lib/env_dpdk/pci_event.o 00:01:39.236 CC lib/env_dpdk/pci_dpdk.o 00:01:39.236 CC lib/env_dpdk/pci_dpdk_2207.o 00:01:39.236 CC lib/env_dpdk/pci_dpdk_2211.o 00:01:39.496 LIB libspdk_conf.a 00:01:39.496 SO libspdk_conf.so.6.0 00:01:39.496 LIB libspdk_rdma_utils.a 00:01:39.496 LIB libspdk_json.a 00:01:39.496 SO libspdk_rdma_utils.so.1.0 00:01:39.496 SYMLINK libspdk_conf.so 00:01:39.496 SO libspdk_json.so.6.0 00:01:39.496 SYMLINK libspdk_rdma_utils.so 00:01:39.496 SYMLINK libspdk_json.so 00:01:39.756 LIB libspdk_idxd.a 00:01:39.756 LIB libspdk_vmd.a 00:01:39.756 SO libspdk_idxd.so.12.1 00:01:39.756 SO libspdk_vmd.so.6.0 00:01:39.756 SYMLINK libspdk_idxd.so 00:01:39.756 SYMLINK libspdk_vmd.so 00:01:40.017 CC lib/rdma_provider/common.o 00:01:40.017 CC lib/rdma_provider/rdma_provider_verbs.o 00:01:40.017 CC lib/jsonrpc/jsonrpc_server.o 00:01:40.017 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:01:40.017 CC lib/jsonrpc/jsonrpc_client.o 00:01:40.017 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:01:40.278 LIB libspdk_rdma_provider.a 00:01:40.278 SO libspdk_rdma_provider.so.7.0 00:01:40.278 LIB libspdk_jsonrpc.a 00:01:40.278 SO libspdk_jsonrpc.so.6.0 00:01:40.278 SYMLINK libspdk_rdma_provider.so 00:01:40.278 SYMLINK libspdk_jsonrpc.so 00:01:40.539 LIB libspdk_env_dpdk.a 00:01:40.539 SO libspdk_env_dpdk.so.15.0 00:01:40.539 SYMLINK libspdk_env_dpdk.so 00:01:40.799 CC lib/rpc/rpc.o 00:01:41.060 LIB libspdk_rpc.a 00:01:41.060 SO libspdk_rpc.so.6.0 00:01:41.060 SYMLINK libspdk_rpc.so 00:01:41.321 CC lib/trace/trace.o 00:01:41.321 CC lib/trace/trace_flags.o 00:01:41.321 CC lib/trace/trace_rpc.o 00:01:41.321 CC lib/notify/notify.o 00:01:41.321 CC lib/notify/notify_rpc.o 00:01:41.321 CC lib/keyring/keyring.o 00:01:41.321 CC lib/keyring/keyring_rpc.o 00:01:41.582 LIB libspdk_notify.a 00:01:41.582 SO libspdk_notify.so.6.0 00:01:41.582 LIB libspdk_trace.a 00:01:41.582 LIB libspdk_keyring.a 00:01:41.582 SO libspdk_trace.so.11.0 00:01:41.582 SO libspdk_keyring.so.2.0 00:01:41.582 SYMLINK libspdk_notify.so 00:01:41.843 SYMLINK libspdk_keyring.so 00:01:41.843 SYMLINK libspdk_trace.so 00:01:42.104 CC lib/thread/thread.o 00:01:42.104 CC lib/thread/iobuf.o 00:01:42.104 CC lib/sock/sock.o 00:01:42.104 CC lib/sock/sock_rpc.o 00:01:42.675 LIB libspdk_sock.a 00:01:42.675 SO libspdk_sock.so.10.0 00:01:42.675 SYMLINK libspdk_sock.so 00:01:42.936 CC lib/nvme/nvme_ctrlr_cmd.o 00:01:42.936 CC lib/nvme/nvme_ctrlr.o 00:01:42.936 CC lib/nvme/nvme_fabric.o 00:01:42.936 CC lib/nvme/nvme_ns_cmd.o 00:01:42.936 CC lib/nvme/nvme_ns.o 00:01:42.936 CC lib/nvme/nvme_pcie_common.o 00:01:42.936 CC lib/nvme/nvme_pcie.o 00:01:42.936 CC lib/nvme/nvme_qpair.o 00:01:42.936 CC lib/nvme/nvme.o 00:01:42.936 CC lib/nvme/nvme_quirks.o 00:01:42.936 CC lib/nvme/nvme_transport.o 00:01:42.936 CC lib/nvme/nvme_discovery.o 00:01:42.936 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:01:42.936 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:01:42.936 CC lib/nvme/nvme_tcp.o 00:01:42.936 CC lib/nvme/nvme_opal.o 00:01:42.936 CC lib/nvme/nvme_io_msg.o 00:01:42.936 CC lib/nvme/nvme_poll_group.o 00:01:42.936 CC lib/nvme/nvme_zns.o 00:01:42.936 CC lib/nvme/nvme_stubs.o 00:01:42.936 CC lib/nvme/nvme_auth.o 00:01:42.936 CC lib/nvme/nvme_cuse.o 00:01:42.936 CC lib/nvme/nvme_vfio_user.o 00:01:42.936 CC lib/nvme/nvme_rdma.o 00:01:43.508 LIB libspdk_thread.a 00:01:43.508 SO libspdk_thread.so.10.2 00:01:43.508 SYMLINK libspdk_thread.so 00:01:43.833 CC lib/virtio/virtio.o 00:01:43.833 CC lib/virtio/virtio_vhost_user.o 00:01:43.833 CC lib/virtio/virtio_vfio_user.o 00:01:43.833 CC lib/virtio/virtio_pci.o 00:01:43.833 CC lib/vfu_tgt/tgt_endpoint.o 00:01:43.833 CC lib/vfu_tgt/tgt_rpc.o 00:01:44.153 CC lib/accel/accel.o 00:01:44.153 CC lib/accel/accel_rpc.o 00:01:44.153 CC lib/accel/accel_sw.o 00:01:44.153 CC lib/init/json_config.o 00:01:44.153 CC lib/init/subsystem_rpc.o 00:01:44.153 CC lib/init/subsystem.o 00:01:44.153 CC lib/init/rpc.o 00:01:44.153 CC lib/fsdev/fsdev.o 00:01:44.153 CC lib/fsdev/fsdev_rpc.o 00:01:44.153 CC lib/fsdev/fsdev_io.o 00:01:44.153 CC lib/blob/blobstore.o 00:01:44.153 CC lib/blob/request.o 00:01:44.153 CC lib/blob/zeroes.o 00:01:44.153 CC lib/blob/blob_bs_dev.o 00:01:44.153 LIB libspdk_init.a 00:01:44.153 SO libspdk_init.so.6.0 00:01:44.461 LIB libspdk_virtio.a 00:01:44.461 LIB libspdk_vfu_tgt.a 00:01:44.461 SO libspdk_virtio.so.7.0 00:01:44.461 SO libspdk_vfu_tgt.so.3.0 00:01:44.461 SYMLINK libspdk_init.so 00:01:44.461 SYMLINK libspdk_vfu_tgt.so 00:01:44.461 SYMLINK libspdk_virtio.so 00:01:44.461 LIB libspdk_fsdev.a 00:01:44.722 SO libspdk_fsdev.so.1.0 00:01:44.722 SYMLINK libspdk_fsdev.so 00:01:44.722 CC lib/event/app.o 00:01:44.722 CC lib/event/reactor.o 00:01:44.722 CC lib/event/log_rpc.o 00:01:44.722 CC lib/event/app_rpc.o 00:01:44.722 CC lib/event/scheduler_static.o 00:01:44.984 LIB libspdk_accel.a 00:01:44.984 LIB libspdk_nvme.a 00:01:44.984 SO libspdk_accel.so.16.0 00:01:44.984 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:01:44.984 SO libspdk_nvme.so.15.0 00:01:44.984 SYMLINK libspdk_accel.so 00:01:45.246 LIB libspdk_event.a 00:01:45.246 SO libspdk_event.so.14.0 00:01:45.246 SYMLINK libspdk_event.so 00:01:45.246 SYMLINK libspdk_nvme.so 00:01:45.508 CC lib/bdev/bdev.o 00:01:45.508 CC lib/bdev/bdev_rpc.o 00:01:45.508 CC lib/bdev/bdev_zone.o 00:01:45.508 CC lib/bdev/part.o 00:01:45.508 CC lib/bdev/scsi_nvme.o 00:01:45.770 LIB libspdk_fuse_dispatcher.a 00:01:45.770 SO libspdk_fuse_dispatcher.so.1.0 00:01:45.770 SYMLINK libspdk_fuse_dispatcher.so 00:01:46.714 LIB libspdk_blob.a 00:01:46.714 SO libspdk_blob.so.11.0 00:01:46.714 SYMLINK libspdk_blob.so 00:01:46.975 CC lib/blobfs/blobfs.o 00:01:46.975 CC lib/lvol/lvol.o 00:01:46.975 CC lib/blobfs/tree.o 00:01:47.919 LIB libspdk_bdev.a 00:01:47.919 SO libspdk_bdev.so.17.0 00:01:47.919 LIB libspdk_blobfs.a 00:01:47.919 SYMLINK libspdk_bdev.so 00:01:47.919 SO libspdk_blobfs.so.10.0 00:01:47.919 LIB libspdk_lvol.a 00:01:47.919 SO libspdk_lvol.so.10.0 00:01:47.919 SYMLINK libspdk_blobfs.so 00:01:47.919 SYMLINK libspdk_lvol.so 00:01:48.180 CC lib/nbd/nbd.o 00:01:48.180 CC lib/nvmf/ctrlr.o 00:01:48.180 CC lib/ublk/ublk.o 00:01:48.180 CC lib/nbd/nbd_rpc.o 00:01:48.180 CC lib/nvmf/ctrlr_discovery.o 00:01:48.180 CC lib/ublk/ublk_rpc.o 00:01:48.180 CC lib/scsi/dev.o 00:01:48.180 CC lib/ftl/ftl_core.o 00:01:48.180 CC lib/nvmf/ctrlr_bdev.o 00:01:48.180 CC lib/scsi/lun.o 00:01:48.180 CC lib/ftl/ftl_init.o 00:01:48.180 CC lib/nvmf/subsystem.o 00:01:48.180 CC lib/scsi/port.o 00:01:48.180 CC lib/scsi/scsi.o 00:01:48.180 CC lib/ftl/ftl_layout.o 00:01:48.180 CC lib/nvmf/nvmf.o 00:01:48.180 CC lib/ftl/ftl_debug.o 00:01:48.180 CC lib/nvmf/nvmf_rpc.o 00:01:48.180 CC lib/scsi/scsi_bdev.o 00:01:48.180 CC lib/ftl/ftl_io.o 00:01:48.180 CC lib/nvmf/transport.o 00:01:48.180 CC lib/ftl/ftl_sb.o 00:01:48.180 CC lib/scsi/scsi_pr.o 00:01:48.180 CC lib/scsi/scsi_rpc.o 00:01:48.180 CC lib/nvmf/tcp.o 00:01:48.180 CC lib/ftl/ftl_l2p.o 00:01:48.180 CC lib/scsi/task.o 00:01:48.180 CC lib/nvmf/stubs.o 00:01:48.180 CC lib/ftl/ftl_l2p_flat.o 00:01:48.180 CC lib/nvmf/mdns_server.o 00:01:48.180 CC lib/ftl/ftl_nv_cache.o 00:01:48.180 CC lib/ftl/ftl_band.o 00:01:48.180 CC lib/nvmf/rdma.o 00:01:48.180 CC lib/nvmf/vfio_user.o 00:01:48.180 CC lib/ftl/ftl_writer.o 00:01:48.180 CC lib/ftl/ftl_band_ops.o 00:01:48.180 CC lib/nvmf/auth.o 00:01:48.180 CC lib/ftl/ftl_rq.o 00:01:48.180 CC lib/ftl/ftl_reloc.o 00:01:48.180 CC lib/ftl/ftl_l2p_cache.o 00:01:48.180 CC lib/ftl/ftl_p2l.o 00:01:48.180 CC lib/ftl/ftl_p2l_log.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_startup.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_md.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_misc.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_band.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:01:48.180 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:01:48.181 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:01:48.181 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:01:48.181 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:01:48.181 CC lib/ftl/utils/ftl_conf.o 00:01:48.181 CC lib/ftl/utils/ftl_md.o 00:01:48.181 CC lib/ftl/utils/ftl_mempool.o 00:01:48.181 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:01:48.181 CC lib/ftl/utils/ftl_bitmap.o 00:01:48.181 CC lib/ftl/utils/ftl_property.o 00:01:48.181 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:01:48.181 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:01:48.181 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:01:48.181 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:01:48.181 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:01:48.181 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:01:48.442 CC lib/ftl/upgrade/ftl_sb_v3.o 00:01:48.442 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:01:48.442 CC lib/ftl/nvc/ftl_nvc_dev.o 00:01:48.442 CC lib/ftl/upgrade/ftl_sb_v5.o 00:01:48.442 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:01:48.442 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:01:48.442 CC lib/ftl/ftl_trace.o 00:01:48.442 CC lib/ftl/base/ftl_base_dev.o 00:01:48.442 CC lib/ftl/base/ftl_base_bdev.o 00:01:49.014 LIB libspdk_nbd.a 00:01:49.014 SO libspdk_nbd.so.7.0 00:01:49.014 SYMLINK libspdk_nbd.so 00:01:49.276 LIB libspdk_scsi.a 00:01:49.276 SO libspdk_scsi.so.9.0 00:01:49.276 LIB libspdk_ublk.a 00:01:49.276 SYMLINK libspdk_scsi.so 00:01:49.276 SO libspdk_ublk.so.3.0 00:01:49.276 SYMLINK libspdk_ublk.so 00:01:49.542 LIB libspdk_ftl.a 00:01:49.542 SO libspdk_ftl.so.9.0 00:01:49.542 CC lib/iscsi/conn.o 00:01:49.542 CC lib/iscsi/init_grp.o 00:01:49.542 CC lib/vhost/vhost.o 00:01:49.542 CC lib/iscsi/iscsi.o 00:01:49.542 CC lib/vhost/vhost_rpc.o 00:01:49.542 CC lib/iscsi/param.o 00:01:49.542 CC lib/vhost/vhost_scsi.o 00:01:49.542 CC lib/iscsi/portal_grp.o 00:01:49.542 CC lib/iscsi/tgt_node.o 00:01:49.542 CC lib/vhost/vhost_blk.o 00:01:49.542 CC lib/iscsi/iscsi_subsystem.o 00:01:49.542 CC lib/vhost/rte_vhost_user.o 00:01:49.542 CC lib/iscsi/iscsi_rpc.o 00:01:49.542 CC lib/iscsi/task.o 00:01:49.807 SYMLINK libspdk_ftl.so 00:01:50.379 LIB libspdk_nvmf.a 00:01:50.640 SO libspdk_nvmf.so.19.1 00:01:50.640 LIB libspdk_vhost.a 00:01:50.640 SO libspdk_vhost.so.8.0 00:01:50.640 SYMLINK libspdk_nvmf.so 00:01:50.902 SYMLINK libspdk_vhost.so 00:01:50.902 LIB libspdk_iscsi.a 00:01:50.902 SO libspdk_iscsi.so.8.0 00:01:51.164 SYMLINK libspdk_iscsi.so 00:01:51.737 CC module/env_dpdk/env_dpdk_rpc.o 00:01:51.737 CC module/vfu_device/vfu_virtio_blk.o 00:01:51.737 CC module/vfu_device/vfu_virtio.o 00:01:51.737 CC module/vfu_device/vfu_virtio_scsi.o 00:01:51.737 CC module/vfu_device/vfu_virtio_rpc.o 00:01:51.737 CC module/vfu_device/vfu_virtio_fs.o 00:01:51.737 LIB libspdk_env_dpdk_rpc.a 00:01:51.737 CC module/scheduler/dynamic/scheduler_dynamic.o 00:01:51.737 CC module/sock/posix/posix.o 00:01:51.737 CC module/keyring/file/keyring.o 00:01:51.737 CC module/keyring/linux/keyring.o 00:01:51.737 CC module/keyring/file/keyring_rpc.o 00:01:51.737 CC module/keyring/linux/keyring_rpc.o 00:01:51.998 CC module/accel/error/accel_error.o 00:01:51.998 CC module/accel/error/accel_error_rpc.o 00:01:51.998 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:01:51.998 CC module/blob/bdev/blob_bdev.o 00:01:51.998 CC module/fsdev/aio/fsdev_aio.o 00:01:51.998 CC module/fsdev/aio/fsdev_aio_rpc.o 00:01:51.998 CC module/fsdev/aio/linux_aio_mgr.o 00:01:51.998 CC module/accel/dsa/accel_dsa_rpc.o 00:01:51.998 CC module/accel/dsa/accel_dsa.o 00:01:51.998 CC module/scheduler/gscheduler/gscheduler.o 00:01:51.998 CC module/accel/ioat/accel_ioat.o 00:01:51.998 CC module/accel/iaa/accel_iaa.o 00:01:51.998 CC module/accel/ioat/accel_ioat_rpc.o 00:01:51.998 CC module/accel/iaa/accel_iaa_rpc.o 00:01:51.998 SO libspdk_env_dpdk_rpc.so.6.0 00:01:51.998 SYMLINK libspdk_env_dpdk_rpc.so 00:01:51.998 LIB libspdk_keyring_linux.a 00:01:51.998 LIB libspdk_scheduler_gscheduler.a 00:01:51.998 LIB libspdk_keyring_file.a 00:01:51.998 LIB libspdk_scheduler_dpdk_governor.a 00:01:51.998 LIB libspdk_accel_error.a 00:01:51.998 SO libspdk_keyring_linux.so.1.0 00:01:51.998 SO libspdk_keyring_file.so.2.0 00:01:51.998 SO libspdk_scheduler_gscheduler.so.4.0 00:01:51.998 LIB libspdk_accel_ioat.a 00:01:51.998 SO libspdk_scheduler_dpdk_governor.so.4.0 00:01:51.998 LIB libspdk_scheduler_dynamic.a 00:01:52.259 SO libspdk_accel_error.so.2.0 00:01:52.259 LIB libspdk_accel_iaa.a 00:01:52.259 SO libspdk_accel_ioat.so.6.0 00:01:52.259 SO libspdk_scheduler_dynamic.so.4.0 00:01:52.259 SYMLINK libspdk_keyring_linux.so 00:01:52.259 SYMLINK libspdk_keyring_file.so 00:01:52.259 SYMLINK libspdk_scheduler_gscheduler.so 00:01:52.259 LIB libspdk_blob_bdev.a 00:01:52.259 SO libspdk_accel_iaa.so.3.0 00:01:52.259 SYMLINK libspdk_scheduler_dpdk_governor.so 00:01:52.259 LIB libspdk_accel_dsa.a 00:01:52.259 SYMLINK libspdk_accel_error.so 00:01:52.259 SYMLINK libspdk_accel_ioat.so 00:01:52.259 SO libspdk_blob_bdev.so.11.0 00:01:52.259 SYMLINK libspdk_scheduler_dynamic.so 00:01:52.259 SO libspdk_accel_dsa.so.5.0 00:01:52.259 SYMLINK libspdk_accel_iaa.so 00:01:52.259 LIB libspdk_vfu_device.a 00:01:52.259 SYMLINK libspdk_blob_bdev.so 00:01:52.259 SO libspdk_vfu_device.so.3.0 00:01:52.259 SYMLINK libspdk_accel_dsa.so 00:01:52.520 SYMLINK libspdk_vfu_device.so 00:01:52.520 LIB libspdk_fsdev_aio.a 00:01:52.520 SO libspdk_fsdev_aio.so.1.0 00:01:52.520 LIB libspdk_sock_posix.a 00:01:52.520 SO libspdk_sock_posix.so.6.0 00:01:52.520 SYMLINK libspdk_fsdev_aio.so 00:01:52.781 SYMLINK libspdk_sock_posix.so 00:01:52.781 CC module/bdev/delay/vbdev_delay.o 00:01:52.781 CC module/bdev/delay/vbdev_delay_rpc.o 00:01:52.781 CC module/bdev/lvol/vbdev_lvol.o 00:01:52.781 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:01:52.781 CC module/bdev/error/vbdev_error.o 00:01:52.781 CC module/bdev/error/vbdev_error_rpc.o 00:01:52.781 CC module/bdev/nvme/bdev_nvme.o 00:01:52.781 CC module/bdev/nvme/bdev_nvme_rpc.o 00:01:52.781 CC module/bdev/nvme/nvme_rpc.o 00:01:52.781 CC module/bdev/gpt/gpt.o 00:01:52.781 CC module/bdev/split/vbdev_split.o 00:01:52.781 CC module/bdev/nvme/bdev_mdns_client.o 00:01:52.781 CC module/bdev/split/vbdev_split_rpc.o 00:01:52.781 CC module/bdev/gpt/vbdev_gpt.o 00:01:52.781 CC module/bdev/nvme/vbdev_opal_rpc.o 00:01:52.781 CC module/bdev/nvme/vbdev_opal.o 00:01:52.781 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:01:52.781 CC module/bdev/null/bdev_null.o 00:01:52.781 CC module/blobfs/bdev/blobfs_bdev.o 00:01:52.781 CC module/bdev/null/bdev_null_rpc.o 00:01:52.781 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:01:52.781 CC module/bdev/zone_block/vbdev_zone_block.o 00:01:52.781 CC module/bdev/malloc/bdev_malloc.o 00:01:52.781 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:01:52.781 CC module/bdev/malloc/bdev_malloc_rpc.o 00:01:52.781 CC module/bdev/aio/bdev_aio.o 00:01:52.781 CC module/bdev/ftl/bdev_ftl.o 00:01:52.781 CC module/bdev/iscsi/bdev_iscsi.o 00:01:52.781 CC module/bdev/aio/bdev_aio_rpc.o 00:01:52.781 CC module/bdev/ftl/bdev_ftl_rpc.o 00:01:52.781 CC module/bdev/raid/bdev_raid_rpc.o 00:01:52.781 CC module/bdev/virtio/bdev_virtio_scsi.o 00:01:52.781 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:01:52.781 CC module/bdev/raid/bdev_raid.o 00:01:52.781 CC module/bdev/virtio/bdev_virtio_blk.o 00:01:52.781 CC module/bdev/passthru/vbdev_passthru.o 00:01:52.781 CC module/bdev/virtio/bdev_virtio_rpc.o 00:01:52.781 CC module/bdev/raid/raid0.o 00:01:52.781 CC module/bdev/raid/bdev_raid_sb.o 00:01:52.781 CC module/bdev/raid/raid1.o 00:01:52.781 CC module/bdev/raid/concat.o 00:01:52.781 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:01:53.352 LIB libspdk_blobfs_bdev.a 00:01:53.352 LIB libspdk_bdev_gpt.a 00:01:53.352 SO libspdk_blobfs_bdev.so.6.0 00:01:53.352 LIB libspdk_bdev_error.a 00:01:53.352 SO libspdk_bdev_gpt.so.6.0 00:01:53.352 LIB libspdk_bdev_ftl.a 00:01:53.352 LIB libspdk_bdev_split.a 00:01:53.352 LIB libspdk_bdev_delay.a 00:01:53.352 LIB libspdk_bdev_null.a 00:01:53.352 SO libspdk_bdev_split.so.6.0 00:01:53.352 SO libspdk_bdev_error.so.6.0 00:01:53.352 SO libspdk_bdev_ftl.so.6.0 00:01:53.352 LIB libspdk_bdev_aio.a 00:01:53.352 SYMLINK libspdk_blobfs_bdev.so 00:01:53.352 SYMLINK libspdk_bdev_gpt.so 00:01:53.352 SO libspdk_bdev_delay.so.6.0 00:01:53.352 SO libspdk_bdev_null.so.6.0 00:01:53.352 LIB libspdk_bdev_passthru.a 00:01:53.352 LIB libspdk_bdev_zone_block.a 00:01:53.352 SO libspdk_bdev_aio.so.6.0 00:01:53.352 SO libspdk_bdev_zone_block.so.6.0 00:01:53.352 SYMLINK libspdk_bdev_ftl.so 00:01:53.352 SYMLINK libspdk_bdev_error.so 00:01:53.352 SYMLINK libspdk_bdev_split.so 00:01:53.352 LIB libspdk_bdev_malloc.a 00:01:53.352 SO libspdk_bdev_passthru.so.6.0 00:01:53.352 LIB libspdk_bdev_iscsi.a 00:01:53.352 SYMLINK libspdk_bdev_null.so 00:01:53.352 SYMLINK libspdk_bdev_delay.so 00:01:53.353 SO libspdk_bdev_malloc.so.6.0 00:01:53.353 SO libspdk_bdev_iscsi.so.6.0 00:01:53.353 SYMLINK libspdk_bdev_aio.so 00:01:53.353 LIB libspdk_bdev_virtio.a 00:01:53.353 LIB libspdk_bdev_lvol.a 00:01:53.353 SYMLINK libspdk_bdev_zone_block.so 00:01:53.353 SYMLINK libspdk_bdev_passthru.so 00:01:53.614 SO libspdk_bdev_lvol.so.6.0 00:01:53.614 SO libspdk_bdev_virtio.so.6.0 00:01:53.614 SYMLINK libspdk_bdev_malloc.so 00:01:53.614 SYMLINK libspdk_bdev_iscsi.so 00:01:53.614 SYMLINK libspdk_bdev_lvol.so 00:01:53.614 SYMLINK libspdk_bdev_virtio.so 00:01:53.875 LIB libspdk_bdev_raid.a 00:01:54.137 SO libspdk_bdev_raid.so.6.0 00:01:54.137 SYMLINK libspdk_bdev_raid.so 00:01:55.081 LIB libspdk_bdev_nvme.a 00:01:55.081 SO libspdk_bdev_nvme.so.7.0 00:01:55.081 SYMLINK libspdk_bdev_nvme.so 00:01:56.025 CC module/event/subsystems/iobuf/iobuf.o 00:01:56.025 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:01:56.025 CC module/event/subsystems/sock/sock.o 00:01:56.025 CC module/event/subsystems/vmd/vmd.o 00:01:56.025 CC module/event/subsystems/vmd/vmd_rpc.o 00:01:56.025 CC module/event/subsystems/keyring/keyring.o 00:01:56.025 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:01:56.025 CC module/event/subsystems/scheduler/scheduler.o 00:01:56.025 CC module/event/subsystems/fsdev/fsdev.o 00:01:56.025 CC module/event/subsystems/vfu_tgt/vfu_tgt.o 00:01:56.025 LIB libspdk_event_vhost_blk.a 00:01:56.025 LIB libspdk_event_keyring.a 00:01:56.025 LIB libspdk_event_sock.a 00:01:56.025 LIB libspdk_event_iobuf.a 00:01:56.025 LIB libspdk_event_vmd.a 00:01:56.025 LIB libspdk_event_vfu_tgt.a 00:01:56.025 LIB libspdk_event_fsdev.a 00:01:56.025 LIB libspdk_event_scheduler.a 00:01:56.025 SO libspdk_event_keyring.so.1.0 00:01:56.025 SO libspdk_event_vhost_blk.so.3.0 00:01:56.025 SO libspdk_event_iobuf.so.3.0 00:01:56.025 SO libspdk_event_sock.so.5.0 00:01:56.025 SO libspdk_event_vfu_tgt.so.3.0 00:01:56.025 SO libspdk_event_fsdev.so.1.0 00:01:56.025 SO libspdk_event_vmd.so.6.0 00:01:56.025 SO libspdk_event_scheduler.so.4.0 00:01:56.025 SYMLINK libspdk_event_keyring.so 00:01:56.025 SYMLINK libspdk_event_vhost_blk.so 00:01:56.285 SYMLINK libspdk_event_sock.so 00:01:56.285 SYMLINK libspdk_event_vfu_tgt.so 00:01:56.285 SYMLINK libspdk_event_iobuf.so 00:01:56.285 SYMLINK libspdk_event_fsdev.so 00:01:56.285 SYMLINK libspdk_event_scheduler.so 00:01:56.285 SYMLINK libspdk_event_vmd.so 00:01:56.546 CC module/event/subsystems/accel/accel.o 00:01:56.858 LIB libspdk_event_accel.a 00:01:56.858 SO libspdk_event_accel.so.6.0 00:01:56.858 SYMLINK libspdk_event_accel.so 00:01:57.119 CC module/event/subsystems/bdev/bdev.o 00:01:57.380 LIB libspdk_event_bdev.a 00:01:57.380 SO libspdk_event_bdev.so.6.0 00:01:57.380 SYMLINK libspdk_event_bdev.so 00:01:57.950 CC module/event/subsystems/scsi/scsi.o 00:01:57.950 CC module/event/subsystems/nbd/nbd.o 00:01:57.950 CC module/event/subsystems/ublk/ublk.o 00:01:57.950 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:01:57.950 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:01:57.950 LIB libspdk_event_nbd.a 00:01:57.950 LIB libspdk_event_scsi.a 00:01:57.950 LIB libspdk_event_ublk.a 00:01:57.950 SO libspdk_event_nbd.so.6.0 00:01:57.950 SO libspdk_event_scsi.so.6.0 00:01:57.950 SO libspdk_event_ublk.so.3.0 00:01:57.950 LIB libspdk_event_nvmf.a 00:01:57.950 SYMLINK libspdk_event_nbd.so 00:01:58.212 SYMLINK libspdk_event_ublk.so 00:01:58.212 SYMLINK libspdk_event_scsi.so 00:01:58.212 SO libspdk_event_nvmf.so.6.0 00:01:58.212 SYMLINK libspdk_event_nvmf.so 00:01:58.474 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:01:58.474 CC module/event/subsystems/iscsi/iscsi.o 00:01:58.736 LIB libspdk_event_vhost_scsi.a 00:01:58.736 LIB libspdk_event_iscsi.a 00:01:58.736 SO libspdk_event_vhost_scsi.so.3.0 00:01:58.736 SO libspdk_event_iscsi.so.6.0 00:01:58.736 SYMLINK libspdk_event_vhost_scsi.so 00:01:58.736 SYMLINK libspdk_event_iscsi.so 00:01:58.997 SO libspdk.so.6.0 00:01:58.997 SYMLINK libspdk.so 00:01:59.257 CC app/trace_record/trace_record.o 00:01:59.257 CXX app/trace/trace.o 00:01:59.257 CC app/spdk_lspci/spdk_lspci.o 00:01:59.257 TEST_HEADER include/spdk/accel.h 00:01:59.257 TEST_HEADER include/spdk/accel_module.h 00:01:59.257 TEST_HEADER include/spdk/assert.h 00:01:59.257 TEST_HEADER include/spdk/barrier.h 00:01:59.257 TEST_HEADER include/spdk/base64.h 00:01:59.257 CC app/spdk_nvme_identify/identify.o 00:01:59.257 TEST_HEADER include/spdk/bdev.h 00:01:59.257 TEST_HEADER include/spdk/bdev_module.h 00:01:59.257 CC test/rpc_client/rpc_client_test.o 00:01:59.257 TEST_HEADER include/spdk/bdev_zone.h 00:01:59.257 TEST_HEADER include/spdk/bit_array.h 00:01:59.257 CC app/spdk_nvme_perf/perf.o 00:01:59.257 CC app/spdk_nvme_discover/discovery_aer.o 00:01:59.257 TEST_HEADER include/spdk/bit_pool.h 00:01:59.257 CC app/spdk_top/spdk_top.o 00:01:59.257 TEST_HEADER include/spdk/blob_bdev.h 00:01:59.257 TEST_HEADER include/spdk/blobfs_bdev.h 00:01:59.257 TEST_HEADER include/spdk/blobfs.h 00:01:59.257 TEST_HEADER include/spdk/blob.h 00:01:59.257 TEST_HEADER include/spdk/conf.h 00:01:59.257 TEST_HEADER include/spdk/cpuset.h 00:01:59.257 TEST_HEADER include/spdk/config.h 00:01:59.257 TEST_HEADER include/spdk/crc16.h 00:01:59.257 TEST_HEADER include/spdk/crc32.h 00:01:59.257 TEST_HEADER include/spdk/crc64.h 00:01:59.257 TEST_HEADER include/spdk/dif.h 00:01:59.257 TEST_HEADER include/spdk/env_dpdk.h 00:01:59.257 TEST_HEADER include/spdk/endian.h 00:01:59.257 TEST_HEADER include/spdk/dma.h 00:01:59.257 TEST_HEADER include/spdk/env.h 00:01:59.520 TEST_HEADER include/spdk/event.h 00:01:59.520 TEST_HEADER include/spdk/fd_group.h 00:01:59.520 TEST_HEADER include/spdk/fd.h 00:01:59.520 CC examples/interrupt_tgt/interrupt_tgt.o 00:01:59.520 TEST_HEADER include/spdk/file.h 00:01:59.520 TEST_HEADER include/spdk/fsdev.h 00:01:59.520 TEST_HEADER include/spdk/fsdev_module.h 00:01:59.520 TEST_HEADER include/spdk/ftl.h 00:01:59.520 TEST_HEADER include/spdk/fuse_dispatcher.h 00:01:59.520 TEST_HEADER include/spdk/gpt_spec.h 00:01:59.520 TEST_HEADER include/spdk/hexlify.h 00:01:59.520 TEST_HEADER include/spdk/histogram_data.h 00:01:59.520 CC app/nvmf_tgt/nvmf_main.o 00:01:59.520 TEST_HEADER include/spdk/idxd_spec.h 00:01:59.520 TEST_HEADER include/spdk/idxd.h 00:01:59.520 CC app/iscsi_tgt/iscsi_tgt.o 00:01:59.521 TEST_HEADER include/spdk/init.h 00:01:59.521 TEST_HEADER include/spdk/ioat.h 00:01:59.521 TEST_HEADER include/spdk/ioat_spec.h 00:01:59.521 TEST_HEADER include/spdk/iscsi_spec.h 00:01:59.521 CC app/spdk_dd/spdk_dd.o 00:01:59.521 TEST_HEADER include/spdk/jsonrpc.h 00:01:59.521 TEST_HEADER include/spdk/json.h 00:01:59.521 TEST_HEADER include/spdk/keyring.h 00:01:59.521 TEST_HEADER include/spdk/likely.h 00:01:59.521 TEST_HEADER include/spdk/keyring_module.h 00:01:59.521 TEST_HEADER include/spdk/log.h 00:01:59.521 TEST_HEADER include/spdk/memory.h 00:01:59.521 TEST_HEADER include/spdk/md5.h 00:01:59.521 TEST_HEADER include/spdk/lvol.h 00:01:59.521 CC app/spdk_tgt/spdk_tgt.o 00:01:59.521 TEST_HEADER include/spdk/net.h 00:01:59.521 TEST_HEADER include/spdk/mmio.h 00:01:59.521 TEST_HEADER include/spdk/nbd.h 00:01:59.521 TEST_HEADER include/spdk/nvme.h 00:01:59.521 TEST_HEADER include/spdk/notify.h 00:01:59.521 TEST_HEADER include/spdk/nvme_intel.h 00:01:59.521 TEST_HEADER include/spdk/nvme_ocssd.h 00:01:59.521 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:01:59.521 TEST_HEADER include/spdk/nvme_spec.h 00:01:59.521 TEST_HEADER include/spdk/nvme_zns.h 00:01:59.521 TEST_HEADER include/spdk/nvmf_cmd.h 00:01:59.521 TEST_HEADER include/spdk/nvmf.h 00:01:59.521 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:01:59.521 TEST_HEADER include/spdk/nvmf_spec.h 00:01:59.521 TEST_HEADER include/spdk/opal.h 00:01:59.521 TEST_HEADER include/spdk/nvmf_transport.h 00:01:59.521 TEST_HEADER include/spdk/opal_spec.h 00:01:59.521 TEST_HEADER include/spdk/pci_ids.h 00:01:59.521 TEST_HEADER include/spdk/pipe.h 00:01:59.521 TEST_HEADER include/spdk/reduce.h 00:01:59.521 TEST_HEADER include/spdk/queue.h 00:01:59.521 TEST_HEADER include/spdk/rpc.h 00:01:59.521 TEST_HEADER include/spdk/scsi.h 00:01:59.521 TEST_HEADER include/spdk/scheduler.h 00:01:59.521 TEST_HEADER include/spdk/scsi_spec.h 00:01:59.521 TEST_HEADER include/spdk/sock.h 00:01:59.521 TEST_HEADER include/spdk/stdinc.h 00:01:59.521 TEST_HEADER include/spdk/string.h 00:01:59.521 TEST_HEADER include/spdk/thread.h 00:01:59.521 TEST_HEADER include/spdk/trace.h 00:01:59.521 TEST_HEADER include/spdk/trace_parser.h 00:01:59.521 TEST_HEADER include/spdk/tree.h 00:01:59.521 TEST_HEADER include/spdk/ublk.h 00:01:59.521 TEST_HEADER include/spdk/util.h 00:01:59.521 TEST_HEADER include/spdk/vfio_user_pci.h 00:01:59.521 TEST_HEADER include/spdk/uuid.h 00:01:59.521 TEST_HEADER include/spdk/vfio_user_spec.h 00:01:59.521 TEST_HEADER include/spdk/version.h 00:01:59.521 TEST_HEADER include/spdk/vmd.h 00:01:59.521 TEST_HEADER include/spdk/xor.h 00:01:59.521 TEST_HEADER include/spdk/vhost.h 00:01:59.521 CXX test/cpp_headers/accel.o 00:01:59.521 CXX test/cpp_headers/accel_module.o 00:01:59.521 TEST_HEADER include/spdk/zipf.h 00:01:59.521 CXX test/cpp_headers/barrier.o 00:01:59.521 CXX test/cpp_headers/assert.o 00:01:59.521 CXX test/cpp_headers/base64.o 00:01:59.521 CXX test/cpp_headers/bit_array.o 00:01:59.521 CXX test/cpp_headers/bdev.o 00:01:59.521 CXX test/cpp_headers/bdev_module.o 00:01:59.521 CXX test/cpp_headers/bdev_zone.o 00:01:59.521 CXX test/cpp_headers/blob_bdev.o 00:01:59.521 CXX test/cpp_headers/bit_pool.o 00:01:59.521 CXX test/cpp_headers/blobfs_bdev.o 00:01:59.521 CXX test/cpp_headers/blob.o 00:01:59.521 CXX test/cpp_headers/blobfs.o 00:01:59.521 CXX test/cpp_headers/cpuset.o 00:01:59.521 CXX test/cpp_headers/conf.o 00:01:59.521 CXX test/cpp_headers/config.o 00:01:59.521 CXX test/cpp_headers/crc64.o 00:01:59.521 CXX test/cpp_headers/crc16.o 00:01:59.521 CXX test/cpp_headers/dif.o 00:01:59.521 CXX test/cpp_headers/endian.o 00:01:59.521 CXX test/cpp_headers/crc32.o 00:01:59.521 CXX test/cpp_headers/dma.o 00:01:59.521 CXX test/cpp_headers/event.o 00:01:59.521 CXX test/cpp_headers/env_dpdk.o 00:01:59.521 CXX test/cpp_headers/fd.o 00:01:59.521 CXX test/cpp_headers/env.o 00:01:59.521 CXX test/cpp_headers/fsdev.o 00:01:59.521 CXX test/cpp_headers/fd_group.o 00:01:59.521 CC examples/util/zipf/zipf.o 00:01:59.521 CXX test/cpp_headers/ftl.o 00:01:59.521 CXX test/cpp_headers/file.o 00:01:59.521 CC test/env/memory/memory_ut.o 00:01:59.521 CXX test/cpp_headers/hexlify.o 00:01:59.521 CC test/thread/poller_perf/poller_perf.o 00:01:59.521 CXX test/cpp_headers/fsdev_module.o 00:01:59.521 CXX test/cpp_headers/gpt_spec.o 00:01:59.521 CXX test/cpp_headers/idxd.o 00:01:59.521 CXX test/cpp_headers/histogram_data.o 00:01:59.521 CXX test/cpp_headers/fuse_dispatcher.o 00:01:59.521 CXX test/cpp_headers/idxd_spec.o 00:01:59.521 CC test/env/vtophys/vtophys.o 00:01:59.521 CXX test/cpp_headers/init.o 00:01:59.521 CXX test/cpp_headers/ioat.o 00:01:59.521 CC test/env/pci/pci_ut.o 00:01:59.521 CXX test/cpp_headers/iscsi_spec.o 00:01:59.521 CXX test/cpp_headers/ioat_spec.o 00:01:59.521 CC examples/ioat/perf/perf.o 00:01:59.521 CC examples/ioat/verify/verify.o 00:01:59.521 CXX test/cpp_headers/jsonrpc.o 00:01:59.521 CXX test/cpp_headers/keyring.o 00:01:59.521 CXX test/cpp_headers/keyring_module.o 00:01:59.521 CXX test/cpp_headers/json.o 00:01:59.521 CXX test/cpp_headers/log.o 00:01:59.521 CC test/app/jsoncat/jsoncat.o 00:01:59.521 CXX test/cpp_headers/lvol.o 00:01:59.521 CXX test/cpp_headers/likely.o 00:01:59.521 CXX test/cpp_headers/memory.o 00:01:59.521 CXX test/cpp_headers/mmio.o 00:01:59.521 CXX test/cpp_headers/nbd.o 00:01:59.521 CC test/dma/test_dma/test_dma.o 00:01:59.789 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:01:59.789 CXX test/cpp_headers/md5.o 00:01:59.789 CXX test/cpp_headers/notify.o 00:01:59.789 CC test/app/histogram_perf/histogram_perf.o 00:01:59.789 CXX test/cpp_headers/nvme.o 00:01:59.789 CXX test/cpp_headers/net.o 00:01:59.789 CXX test/cpp_headers/nvme_intel.o 00:01:59.789 CXX test/cpp_headers/nvme_ocssd_spec.o 00:01:59.789 CC app/fio/nvme/fio_plugin.o 00:01:59.789 CC test/app/stub/stub.o 00:01:59.789 CXX test/cpp_headers/nvme_ocssd.o 00:01:59.789 CXX test/cpp_headers/nvmf_cmd.o 00:01:59.789 CXX test/cpp_headers/nvmf_fc_spec.o 00:01:59.789 CXX test/cpp_headers/nvme_spec.o 00:01:59.789 CXX test/cpp_headers/nvme_zns.o 00:01:59.789 CXX test/cpp_headers/nvmf_transport.o 00:01:59.789 CXX test/cpp_headers/opal.o 00:01:59.789 CXX test/cpp_headers/nvmf.o 00:01:59.789 CXX test/cpp_headers/nvmf_spec.o 00:01:59.789 CXX test/cpp_headers/reduce.o 00:01:59.789 CXX test/cpp_headers/queue.o 00:01:59.789 CXX test/cpp_headers/opal_spec.o 00:01:59.789 CXX test/cpp_headers/pci_ids.o 00:01:59.789 CXX test/cpp_headers/pipe.o 00:01:59.789 CXX test/cpp_headers/scsi.o 00:01:59.789 CXX test/cpp_headers/rpc.o 00:01:59.789 LINK spdk_lspci 00:01:59.789 CXX test/cpp_headers/scheduler.o 00:01:59.789 CXX test/cpp_headers/stdinc.o 00:01:59.789 CXX test/cpp_headers/scsi_spec.o 00:01:59.789 CXX test/cpp_headers/sock.o 00:01:59.789 CXX test/cpp_headers/trace.o 00:01:59.789 CXX test/cpp_headers/trace_parser.o 00:01:59.789 CXX test/cpp_headers/string.o 00:01:59.789 CXX test/cpp_headers/thread.o 00:01:59.789 CXX test/cpp_headers/util.o 00:01:59.789 CXX test/cpp_headers/tree.o 00:01:59.789 CXX test/cpp_headers/uuid.o 00:01:59.789 CXX test/cpp_headers/ublk.o 00:01:59.789 LINK iscsi_tgt 00:01:59.789 CXX test/cpp_headers/vfio_user_spec.o 00:01:59.789 CXX test/cpp_headers/version.o 00:01:59.789 CC app/fio/bdev/fio_plugin.o 00:01:59.789 CXX test/cpp_headers/vfio_user_pci.o 00:01:59.789 CXX test/cpp_headers/vhost.o 00:01:59.789 CXX test/cpp_headers/vmd.o 00:01:59.789 CXX test/cpp_headers/xor.o 00:01:59.789 CC test/app/bdev_svc/bdev_svc.o 00:01:59.789 CXX test/cpp_headers/zipf.o 00:02:00.057 LINK spdk_tgt 00:02:00.057 LINK spdk_trace 00:02:00.057 LINK spdk_nvme_discover 00:02:00.321 LINK histogram_perf 00:02:00.321 LINK rpc_client_test 00:02:00.321 LINK interrupt_tgt 00:02:00.321 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:02:00.321 CC test/env/mem_callbacks/mem_callbacks.o 00:02:00.321 LINK spdk_dd 00:02:00.321 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:02:00.321 LINK nvmf_tgt 00:02:00.321 LINK poller_perf 00:02:00.321 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:02:00.321 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:02:00.581 LINK spdk_trace_record 00:02:00.581 LINK verify 00:02:00.581 LINK vtophys 00:02:00.581 LINK stub 00:02:00.581 LINK env_dpdk_post_init 00:02:00.842 LINK spdk_nvme_perf 00:02:01.102 LINK spdk_nvme_identify 00:02:01.102 LINK zipf 00:02:01.102 LINK jsoncat 00:02:01.102 CC test/event/event_perf/event_perf.o 00:02:01.102 CC test/event/reactor_perf/reactor_perf.o 00:02:01.102 CC test/event/reactor/reactor.o 00:02:01.102 LINK pci_ut 00:02:01.102 CC test/event/app_repeat/app_repeat.o 00:02:01.102 LINK nvme_fuzz 00:02:01.102 CC test/event/scheduler/scheduler.o 00:02:01.102 LINK ioat_perf 00:02:01.102 LINK test_dma 00:02:01.364 LINK mem_callbacks 00:02:01.364 LINK bdev_svc 00:02:01.364 LINK reactor 00:02:01.364 LINK reactor_perf 00:02:01.364 LINK event_perf 00:02:01.364 LINK app_repeat 00:02:01.364 CC app/vhost/vhost.o 00:02:01.364 LINK scheduler 00:02:01.364 LINK vhost_fuzz 00:02:01.626 CC examples/vmd/lsvmd/lsvmd.o 00:02:01.626 CC examples/vmd/led/led.o 00:02:01.626 CC examples/idxd/perf/perf.o 00:02:01.626 CC examples/sock/hello_world/hello_sock.o 00:02:01.626 LINK spdk_bdev 00:02:01.626 CC examples/thread/thread/thread_ex.o 00:02:01.626 LINK vhost 00:02:01.626 LINK spdk_nvme 00:02:01.626 LINK lsvmd 00:02:01.626 LINK memory_ut 00:02:01.626 LINK led 00:02:01.887 LINK spdk_top 00:02:01.887 LINK hello_sock 00:02:01.887 CC test/nvme/aer/aer.o 00:02:01.887 CC test/nvme/sgl/sgl.o 00:02:01.887 CC test/nvme/connect_stress/connect_stress.o 00:02:01.887 CC test/nvme/reserve/reserve.o 00:02:01.887 CC test/nvme/overhead/overhead.o 00:02:01.887 CC test/nvme/e2edp/nvme_dp.o 00:02:01.887 CC test/nvme/compliance/nvme_compliance.o 00:02:01.887 CC test/nvme/simple_copy/simple_copy.o 00:02:01.887 CC test/nvme/fdp/fdp.o 00:02:01.887 CC test/nvme/boot_partition/boot_partition.o 00:02:01.887 CC test/nvme/err_injection/err_injection.o 00:02:01.887 CC test/nvme/doorbell_aers/doorbell_aers.o 00:02:01.887 CC test/nvme/startup/startup.o 00:02:01.887 CC test/nvme/reset/reset.o 00:02:01.887 CC test/nvme/fused_ordering/fused_ordering.o 00:02:01.887 CC test/nvme/cuse/cuse.o 00:02:01.887 LINK idxd_perf 00:02:01.887 CC test/accel/dif/dif.o 00:02:01.887 CC test/blobfs/mkfs/mkfs.o 00:02:01.887 LINK thread 00:02:02.149 CC test/lvol/esnap/esnap.o 00:02:02.149 LINK connect_stress 00:02:02.149 LINK err_injection 00:02:02.149 LINK boot_partition 00:02:02.149 LINK startup 00:02:02.149 LINK doorbell_aers 00:02:02.149 LINK reserve 00:02:02.149 LINK fused_ordering 00:02:02.149 LINK mkfs 00:02:02.149 LINK overhead 00:02:02.149 LINK sgl 00:02:02.149 LINK simple_copy 00:02:02.149 LINK aer 00:02:02.149 LINK nvme_dp 00:02:02.149 LINK reset 00:02:02.149 LINK nvme_compliance 00:02:02.149 LINK fdp 00:02:02.410 LINK iscsi_fuzz 00:02:02.410 CC examples/nvme/cmb_copy/cmb_copy.o 00:02:02.410 CC examples/nvme/reconnect/reconnect.o 00:02:02.410 CC examples/nvme/arbitration/arbitration.o 00:02:02.410 CC examples/nvme/nvme_manage/nvme_manage.o 00:02:02.410 CC examples/nvme/hotplug/hotplug.o 00:02:02.410 CC examples/nvme/hello_world/hello_world.o 00:02:02.410 CC examples/nvme/abort/abort.o 00:02:02.410 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:02:02.410 CC examples/accel/perf/accel_perf.o 00:02:02.672 LINK dif 00:02:02.672 CC examples/blob/cli/blobcli.o 00:02:02.672 CC examples/blob/hello_world/hello_blob.o 00:02:02.672 CC examples/fsdev/hello_world/hello_fsdev.o 00:02:02.672 LINK cmb_copy 00:02:02.672 LINK pmr_persistence 00:02:02.672 LINK hello_world 00:02:02.672 LINK hotplug 00:02:02.672 LINK arbitration 00:02:02.672 LINK reconnect 00:02:02.672 LINK abort 00:02:02.934 LINK nvme_manage 00:02:02.934 LINK hello_blob 00:02:02.934 LINK hello_fsdev 00:02:02.934 LINK accel_perf 00:02:03.196 LINK blobcli 00:02:03.196 LINK cuse 00:02:03.196 CC test/bdev/bdevio/bdevio.o 00:02:03.458 LINK bdevio 00:02:03.719 CC examples/bdev/hello_world/hello_bdev.o 00:02:03.719 CC examples/bdev/bdevperf/bdevperf.o 00:02:03.981 LINK hello_bdev 00:02:04.242 LINK bdevperf 00:02:05.207 CC examples/nvmf/nvmf/nvmf.o 00:02:05.207 LINK nvmf 00:02:06.599 LINK esnap 00:02:06.599 00:02:06.599 real 0m56.415s 00:02:06.599 user 8m12.343s 00:02:06.599 sys 6m10.587s 00:02:06.599 17:23:14 make -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:06.599 17:23:14 make -- common/autotest_common.sh@10 -- $ set +x 00:02:06.599 ************************************ 00:02:06.599 END TEST make 00:02:06.599 ************************************ 00:02:06.599 17:23:14 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:02:06.599 17:23:14 -- pm/common@29 -- $ signal_monitor_resources TERM 00:02:06.599 17:23:14 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:02:06.599 17:23:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:06.599 17:23:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:02:06.599 17:23:14 -- pm/common@44 -- $ pid=3911411 00:02:06.599 17:23:14 -- pm/common@50 -- $ kill -TERM 3911411 00:02:06.599 17:23:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:06.599 17:23:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:02:06.599 17:23:14 -- pm/common@44 -- $ pid=3911412 00:02:06.599 17:23:14 -- pm/common@50 -- $ kill -TERM 3911412 00:02:06.599 17:23:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:06.599 17:23:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:02:06.599 17:23:14 -- pm/common@44 -- $ pid=3911414 00:02:06.599 17:23:14 -- pm/common@50 -- $ kill -TERM 3911414 00:02:06.599 17:23:14 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:06.599 17:23:14 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:02:06.599 17:23:14 -- pm/common@44 -- $ pid=3911437 00:02:06.599 17:23:14 -- pm/common@50 -- $ sudo -E kill -TERM 3911437 00:02:06.861 17:23:14 -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:02:06.861 17:23:14 -- common/autotest_common.sh@1691 -- # lcov --version 00:02:06.861 17:23:14 -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:02:06.861 17:23:14 -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:02:06.861 17:23:14 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:02:06.861 17:23:14 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:02:06.861 17:23:14 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:02:06.861 17:23:14 -- scripts/common.sh@336 -- # IFS=.-: 00:02:06.861 17:23:14 -- scripts/common.sh@336 -- # read -ra ver1 00:02:06.861 17:23:14 -- scripts/common.sh@337 -- # IFS=.-: 00:02:06.861 17:23:14 -- scripts/common.sh@337 -- # read -ra ver2 00:02:06.861 17:23:14 -- scripts/common.sh@338 -- # local 'op=<' 00:02:06.861 17:23:14 -- scripts/common.sh@340 -- # ver1_l=2 00:02:06.861 17:23:14 -- scripts/common.sh@341 -- # ver2_l=1 00:02:06.861 17:23:14 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:02:06.861 17:23:14 -- scripts/common.sh@344 -- # case "$op" in 00:02:06.861 17:23:14 -- scripts/common.sh@345 -- # : 1 00:02:06.861 17:23:14 -- scripts/common.sh@364 -- # (( v = 0 )) 00:02:06.861 17:23:14 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:06.861 17:23:14 -- scripts/common.sh@365 -- # decimal 1 00:02:06.861 17:23:14 -- scripts/common.sh@353 -- # local d=1 00:02:06.861 17:23:14 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:02:06.861 17:23:14 -- scripts/common.sh@355 -- # echo 1 00:02:06.861 17:23:14 -- scripts/common.sh@365 -- # ver1[v]=1 00:02:06.861 17:23:14 -- scripts/common.sh@366 -- # decimal 2 00:02:06.861 17:23:14 -- scripts/common.sh@353 -- # local d=2 00:02:06.861 17:23:14 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:02:06.861 17:23:14 -- scripts/common.sh@355 -- # echo 2 00:02:06.861 17:23:14 -- scripts/common.sh@366 -- # ver2[v]=2 00:02:06.861 17:23:14 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:02:06.861 17:23:14 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:02:06.861 17:23:14 -- scripts/common.sh@368 -- # return 0 00:02:06.861 17:23:14 -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:02:06.861 17:23:14 -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:02:06.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:06.861 --rc genhtml_branch_coverage=1 00:02:06.861 --rc genhtml_function_coverage=1 00:02:06.861 --rc genhtml_legend=1 00:02:06.861 --rc geninfo_all_blocks=1 00:02:06.861 --rc geninfo_unexecuted_blocks=1 00:02:06.861 00:02:06.861 ' 00:02:06.861 17:23:14 -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:02:06.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:06.861 --rc genhtml_branch_coverage=1 00:02:06.861 --rc genhtml_function_coverage=1 00:02:06.861 --rc genhtml_legend=1 00:02:06.861 --rc geninfo_all_blocks=1 00:02:06.861 --rc geninfo_unexecuted_blocks=1 00:02:06.861 00:02:06.861 ' 00:02:06.861 17:23:14 -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:02:06.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:06.861 --rc genhtml_branch_coverage=1 00:02:06.861 --rc genhtml_function_coverage=1 00:02:06.861 --rc genhtml_legend=1 00:02:06.861 --rc geninfo_all_blocks=1 00:02:06.861 --rc geninfo_unexecuted_blocks=1 00:02:06.861 00:02:06.861 ' 00:02:06.861 17:23:14 -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:02:06.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:02:06.861 --rc genhtml_branch_coverage=1 00:02:06.861 --rc genhtml_function_coverage=1 00:02:06.861 --rc genhtml_legend=1 00:02:06.861 --rc geninfo_all_blocks=1 00:02:06.861 --rc geninfo_unexecuted_blocks=1 00:02:06.861 00:02:06.861 ' 00:02:06.861 17:23:14 -- spdk/autotest.sh@25 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:02:06.861 17:23:14 -- nvmf/common.sh@7 -- # uname -s 00:02:06.861 17:23:14 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:02:06.861 17:23:14 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:02:06.861 17:23:14 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:02:06.861 17:23:14 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:02:06.861 17:23:14 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:02:06.861 17:23:14 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:02:06.861 17:23:14 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:02:06.861 17:23:14 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:02:06.861 17:23:14 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:02:06.861 17:23:14 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:02:06.861 17:23:14 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:02:06.861 17:23:14 -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:02:06.861 17:23:14 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:02:06.861 17:23:14 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:02:06.861 17:23:14 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:02:06.861 17:23:14 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:02:06.861 17:23:14 -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:02:06.861 17:23:14 -- scripts/common.sh@15 -- # shopt -s extglob 00:02:06.861 17:23:14 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:02:06.861 17:23:14 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:06.861 17:23:14 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:06.861 17:23:14 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:06.861 17:23:14 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:06.861 17:23:14 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:06.861 17:23:14 -- paths/export.sh@5 -- # export PATH 00:02:06.861 17:23:14 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:06.861 17:23:14 -- nvmf/common.sh@51 -- # : 0 00:02:06.861 17:23:14 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:02:06.861 17:23:14 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:02:06.861 17:23:14 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:02:06.861 17:23:14 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:02:06.861 17:23:14 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:02:06.861 17:23:14 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:02:06.861 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:02:06.861 17:23:14 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:02:06.861 17:23:14 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:02:06.861 17:23:14 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:02:06.861 17:23:14 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:02:06.862 17:23:14 -- spdk/autotest.sh@32 -- # uname -s 00:02:06.862 17:23:14 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:02:06.862 17:23:14 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:02:06.862 17:23:14 -- spdk/autotest.sh@34 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:02:06.862 17:23:14 -- spdk/autotest.sh@39 -- # echo '|/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/core-collector.sh %P %s %t' 00:02:06.862 17:23:14 -- spdk/autotest.sh@40 -- # echo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:02:06.862 17:23:14 -- spdk/autotest.sh@44 -- # modprobe nbd 00:02:07.122 17:23:14 -- spdk/autotest.sh@46 -- # type -P udevadm 00:02:07.122 17:23:14 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:02:07.122 17:23:14 -- spdk/autotest.sh@48 -- # udevadm_pid=3976987 00:02:07.122 17:23:14 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:02:07.122 17:23:14 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:02:07.122 17:23:14 -- pm/common@17 -- # local monitor 00:02:07.122 17:23:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:07.122 17:23:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:07.122 17:23:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:07.122 17:23:14 -- pm/common@21 -- # date +%s 00:02:07.122 17:23:14 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:07.122 17:23:14 -- pm/common@25 -- # sleep 1 00:02:07.122 17:23:14 -- pm/common@21 -- # date +%s 00:02:07.122 17:23:14 -- pm/common@21 -- # date +%s 00:02:07.122 17:23:14 -- pm/common@21 -- # date +%s 00:02:07.122 17:23:14 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1729178594 00:02:07.122 17:23:14 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1729178594 00:02:07.122 17:23:14 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1729178594 00:02:07.122 17:23:14 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1729178594 00:02:07.122 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1729178594_collect-cpu-load.pm.log 00:02:07.122 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1729178594_collect-vmstat.pm.log 00:02:07.123 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1729178594_collect-cpu-temp.pm.log 00:02:07.123 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1729178594_collect-bmc-pm.bmc.pm.log 00:02:08.066 17:23:15 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:02:08.066 17:23:15 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:02:08.066 17:23:15 -- common/autotest_common.sh@724 -- # xtrace_disable 00:02:08.066 17:23:15 -- common/autotest_common.sh@10 -- # set +x 00:02:08.066 17:23:15 -- spdk/autotest.sh@59 -- # create_test_list 00:02:08.066 17:23:15 -- common/autotest_common.sh@748 -- # xtrace_disable 00:02:08.066 17:23:15 -- common/autotest_common.sh@10 -- # set +x 00:02:08.066 17:23:15 -- spdk/autotest.sh@61 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh 00:02:08.066 17:23:15 -- spdk/autotest.sh@61 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:08.066 17:23:15 -- spdk/autotest.sh@61 -- # src=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:08.066 17:23:15 -- spdk/autotest.sh@62 -- # out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:02:08.066 17:23:15 -- spdk/autotest.sh@63 -- # cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:08.066 17:23:15 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:02:08.066 17:23:15 -- common/autotest_common.sh@1455 -- # uname 00:02:08.066 17:23:15 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:02:08.066 17:23:15 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:02:08.066 17:23:15 -- common/autotest_common.sh@1475 -- # uname 00:02:08.066 17:23:15 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:02:08.066 17:23:15 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:02:08.066 17:23:15 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:02:08.066 lcov: LCOV version 1.15 00:02:08.066 17:23:15 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info 00:02:22.976 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:02:22.976 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno 00:02:41.100 17:23:46 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:02:41.100 17:23:46 -- common/autotest_common.sh@724 -- # xtrace_disable 00:02:41.100 17:23:46 -- common/autotest_common.sh@10 -- # set +x 00:02:41.100 17:23:46 -- spdk/autotest.sh@78 -- # rm -f 00:02:41.100 17:23:46 -- spdk/autotest.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:02:42.044 0000:80:01.6 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:80:01.7 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:80:01.4 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:80:01.5 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:80:01.2 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:80:01.3 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:80:01.0 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:80:01.1 (8086 0b00): Already using the ioatdma driver 00:02:42.044 0000:65:00.0 (144d a80a): Already using the nvme driver 00:02:42.304 0000:00:01.6 (8086 0b00): Already using the ioatdma driver 00:02:42.304 0000:00:01.7 (8086 0b00): Already using the ioatdma driver 00:02:42.304 0000:00:01.4 (8086 0b00): Already using the ioatdma driver 00:02:42.304 0000:00:01.5 (8086 0b00): Already using the ioatdma driver 00:02:42.304 0000:00:01.2 (8086 0b00): Already using the ioatdma driver 00:02:42.304 0000:00:01.3 (8086 0b00): Already using the ioatdma driver 00:02:42.304 0000:00:01.0 (8086 0b00): Already using the ioatdma driver 00:02:42.304 0000:00:01.1 (8086 0b00): Already using the ioatdma driver 00:02:42.565 17:23:50 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:02:42.565 17:23:50 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:02:42.565 17:23:50 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:02:42.565 17:23:50 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:02:42.565 17:23:50 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:02:42.565 17:23:50 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:02:42.565 17:23:50 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:02:42.565 17:23:50 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:02:42.565 17:23:50 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:02:42.565 17:23:50 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:02:42.565 17:23:50 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:02:42.565 17:23:50 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:02:42.565 17:23:50 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:02:42.565 17:23:50 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:02:42.565 17:23:50 -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:02:42.827 No valid GPT data, bailing 00:02:42.827 17:23:50 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:02:42.827 17:23:50 -- scripts/common.sh@394 -- # pt= 00:02:42.827 17:23:50 -- scripts/common.sh@395 -- # return 1 00:02:42.827 17:23:50 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:02:42.827 1+0 records in 00:02:42.827 1+0 records out 00:02:42.827 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00487865 s, 215 MB/s 00:02:42.827 17:23:50 -- spdk/autotest.sh@105 -- # sync 00:02:42.827 17:23:50 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:02:42.827 17:23:50 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:02:42.827 17:23:50 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:02:52.836 17:23:59 -- spdk/autotest.sh@111 -- # uname -s 00:02:52.836 17:23:59 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:02:52.836 17:23:59 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:02:52.836 17:23:59 -- spdk/autotest.sh@115 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:02:54.753 Hugepages 00:02:54.753 node hugesize free / total 00:02:54.753 node0 1048576kB 0 / 0 00:02:54.753 node0 2048kB 0 / 0 00:02:54.753 node1 1048576kB 0 / 0 00:02:54.753 node1 2048kB 0 / 0 00:02:54.753 00:02:54.753 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:54.754 I/OAT 0000:00:01.0 8086 0b00 0 ioatdma - - 00:02:54.754 I/OAT 0000:00:01.1 8086 0b00 0 ioatdma - - 00:02:54.754 I/OAT 0000:00:01.2 8086 0b00 0 ioatdma - - 00:02:54.754 I/OAT 0000:00:01.3 8086 0b00 0 ioatdma - - 00:02:54.754 I/OAT 0000:00:01.4 8086 0b00 0 ioatdma - - 00:02:54.754 I/OAT 0000:00:01.5 8086 0b00 0 ioatdma - - 00:02:54.754 I/OAT 0000:00:01.6 8086 0b00 0 ioatdma - - 00:02:54.754 I/OAT 0000:00:01.7 8086 0b00 0 ioatdma - - 00:02:55.015 NVMe 0000:65:00.0 144d a80a 0 nvme nvme0 nvme0n1 00:02:55.015 I/OAT 0000:80:01.0 8086 0b00 1 ioatdma - - 00:02:55.015 I/OAT 0000:80:01.1 8086 0b00 1 ioatdma - - 00:02:55.015 I/OAT 0000:80:01.2 8086 0b00 1 ioatdma - - 00:02:55.015 I/OAT 0000:80:01.3 8086 0b00 1 ioatdma - - 00:02:55.015 I/OAT 0000:80:01.4 8086 0b00 1 ioatdma - - 00:02:55.015 I/OAT 0000:80:01.5 8086 0b00 1 ioatdma - - 00:02:55.015 I/OAT 0000:80:01.6 8086 0b00 1 ioatdma - - 00:02:55.015 I/OAT 0000:80:01.7 8086 0b00 1 ioatdma - - 00:02:55.015 17:24:02 -- spdk/autotest.sh@117 -- # uname -s 00:02:55.015 17:24:02 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:02:55.015 17:24:02 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:02:55.015 17:24:02 -- common/autotest_common.sh@1514 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:02:59.227 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:02:59.227 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:02:59.227 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:02:59.228 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:03:00.618 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:03:00.879 17:24:08 -- common/autotest_common.sh@1515 -- # sleep 1 00:03:01.824 17:24:09 -- common/autotest_common.sh@1516 -- # bdfs=() 00:03:01.824 17:24:09 -- common/autotest_common.sh@1516 -- # local bdfs 00:03:01.824 17:24:09 -- common/autotest_common.sh@1518 -- # bdfs=($(get_nvme_bdfs)) 00:03:01.824 17:24:09 -- common/autotest_common.sh@1518 -- # get_nvme_bdfs 00:03:01.824 17:24:09 -- common/autotest_common.sh@1496 -- # bdfs=() 00:03:01.824 17:24:09 -- common/autotest_common.sh@1496 -- # local bdfs 00:03:01.824 17:24:09 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:03:01.824 17:24:09 -- common/autotest_common.sh@1497 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:03:01.824 17:24:09 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:03:01.824 17:24:09 -- common/autotest_common.sh@1498 -- # (( 1 == 0 )) 00:03:01.824 17:24:09 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:65:00.0 00:03:01.824 17:24:09 -- common/autotest_common.sh@1520 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:03:06.035 Waiting for block devices as requested 00:03:06.035 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:03:06.035 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:03:06.297 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:03:06.297 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:03:06.297 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:03:06.559 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:03:06.559 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:03:06.559 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:03:06.820 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:03:06.820 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:03:07.082 17:24:14 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:03:07.082 17:24:14 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:65:00.0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1485 -- # grep 0000:65:00.0/nvme/nvme 00:03:07.082 17:24:14 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 ]] 00:03:07.082 17:24:14 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme0 ]] 00:03:07.082 17:24:14 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1529 -- # grep oacs 00:03:07.082 17:24:14 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:03:07.082 17:24:14 -- common/autotest_common.sh@1529 -- # oacs=' 0x5f' 00:03:07.082 17:24:14 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:03:07.082 17:24:14 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:03:07.082 17:24:14 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme0 00:03:07.082 17:24:14 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:03:07.082 17:24:14 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:03:07.082 17:24:14 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:03:07.082 17:24:14 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:03:07.082 17:24:14 -- common/autotest_common.sh@1541 -- # continue 00:03:07.082 17:24:14 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:03:07.082 17:24:14 -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:07.082 17:24:14 -- common/autotest_common.sh@10 -- # set +x 00:03:07.082 17:24:14 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:03:07.082 17:24:14 -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:07.082 17:24:14 -- common/autotest_common.sh@10 -- # set +x 00:03:07.082 17:24:14 -- spdk/autotest.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:11.291 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:03:11.291 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:03:11.291 17:24:19 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:03:11.291 17:24:19 -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:11.291 17:24:19 -- common/autotest_common.sh@10 -- # set +x 00:03:11.291 17:24:19 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:03:11.291 17:24:19 -- common/autotest_common.sh@1576 -- # mapfile -t bdfs 00:03:11.291 17:24:19 -- common/autotest_common.sh@1576 -- # get_nvme_bdfs_by_id 0x0a54 00:03:11.291 17:24:19 -- common/autotest_common.sh@1561 -- # bdfs=() 00:03:11.291 17:24:19 -- common/autotest_common.sh@1561 -- # _bdfs=() 00:03:11.291 17:24:19 -- common/autotest_common.sh@1561 -- # local bdfs _bdfs 00:03:11.291 17:24:19 -- common/autotest_common.sh@1562 -- # _bdfs=($(get_nvme_bdfs)) 00:03:11.291 17:24:19 -- common/autotest_common.sh@1562 -- # get_nvme_bdfs 00:03:11.291 17:24:19 -- common/autotest_common.sh@1496 -- # bdfs=() 00:03:11.291 17:24:19 -- common/autotest_common.sh@1496 -- # local bdfs 00:03:11.291 17:24:19 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:03:11.291 17:24:19 -- common/autotest_common.sh@1497 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:03:11.291 17:24:19 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:03:11.291 17:24:19 -- common/autotest_common.sh@1498 -- # (( 1 == 0 )) 00:03:11.291 17:24:19 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:65:00.0 00:03:11.291 17:24:19 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:03:11.291 17:24:19 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:65:00.0/device 00:03:11.291 17:24:19 -- common/autotest_common.sh@1564 -- # device=0xa80a 00:03:11.291 17:24:19 -- common/autotest_common.sh@1565 -- # [[ 0xa80a == \0\x\0\a\5\4 ]] 00:03:11.291 17:24:19 -- common/autotest_common.sh@1570 -- # (( 0 > 0 )) 00:03:11.291 17:24:19 -- common/autotest_common.sh@1570 -- # return 0 00:03:11.291 17:24:19 -- common/autotest_common.sh@1577 -- # [[ -z '' ]] 00:03:11.291 17:24:19 -- common/autotest_common.sh@1578 -- # return 0 00:03:11.291 17:24:19 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:03:11.291 17:24:19 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:03:11.291 17:24:19 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:03:11.291 17:24:19 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:03:11.291 17:24:19 -- spdk/autotest.sh@149 -- # timing_enter lib 00:03:11.291 17:24:19 -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:11.291 17:24:19 -- common/autotest_common.sh@10 -- # set +x 00:03:11.291 17:24:19 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:03:11.291 17:24:19 -- spdk/autotest.sh@155 -- # run_test env /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:03:11.291 17:24:19 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:11.291 17:24:19 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:11.291 17:24:19 -- common/autotest_common.sh@10 -- # set +x 00:03:11.552 ************************************ 00:03:11.552 START TEST env 00:03:11.552 ************************************ 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:03:11.552 * Looking for test storage... 00:03:11.552 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1691 -- # lcov --version 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:11.552 17:24:19 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:11.552 17:24:19 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:11.552 17:24:19 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:11.552 17:24:19 env -- scripts/common.sh@336 -- # IFS=.-: 00:03:11.552 17:24:19 env -- scripts/common.sh@336 -- # read -ra ver1 00:03:11.552 17:24:19 env -- scripts/common.sh@337 -- # IFS=.-: 00:03:11.552 17:24:19 env -- scripts/common.sh@337 -- # read -ra ver2 00:03:11.552 17:24:19 env -- scripts/common.sh@338 -- # local 'op=<' 00:03:11.552 17:24:19 env -- scripts/common.sh@340 -- # ver1_l=2 00:03:11.552 17:24:19 env -- scripts/common.sh@341 -- # ver2_l=1 00:03:11.552 17:24:19 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:11.552 17:24:19 env -- scripts/common.sh@344 -- # case "$op" in 00:03:11.552 17:24:19 env -- scripts/common.sh@345 -- # : 1 00:03:11.552 17:24:19 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:11.552 17:24:19 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:11.552 17:24:19 env -- scripts/common.sh@365 -- # decimal 1 00:03:11.552 17:24:19 env -- scripts/common.sh@353 -- # local d=1 00:03:11.552 17:24:19 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:11.552 17:24:19 env -- scripts/common.sh@355 -- # echo 1 00:03:11.552 17:24:19 env -- scripts/common.sh@365 -- # ver1[v]=1 00:03:11.552 17:24:19 env -- scripts/common.sh@366 -- # decimal 2 00:03:11.552 17:24:19 env -- scripts/common.sh@353 -- # local d=2 00:03:11.552 17:24:19 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:11.552 17:24:19 env -- scripts/common.sh@355 -- # echo 2 00:03:11.552 17:24:19 env -- scripts/common.sh@366 -- # ver2[v]=2 00:03:11.552 17:24:19 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:11.552 17:24:19 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:11.552 17:24:19 env -- scripts/common.sh@368 -- # return 0 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:11.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:11.552 --rc genhtml_branch_coverage=1 00:03:11.552 --rc genhtml_function_coverage=1 00:03:11.552 --rc genhtml_legend=1 00:03:11.552 --rc geninfo_all_blocks=1 00:03:11.552 --rc geninfo_unexecuted_blocks=1 00:03:11.552 00:03:11.552 ' 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:11.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:11.552 --rc genhtml_branch_coverage=1 00:03:11.552 --rc genhtml_function_coverage=1 00:03:11.552 --rc genhtml_legend=1 00:03:11.552 --rc geninfo_all_blocks=1 00:03:11.552 --rc geninfo_unexecuted_blocks=1 00:03:11.552 00:03:11.552 ' 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:11.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:11.552 --rc genhtml_branch_coverage=1 00:03:11.552 --rc genhtml_function_coverage=1 00:03:11.552 --rc genhtml_legend=1 00:03:11.552 --rc geninfo_all_blocks=1 00:03:11.552 --rc geninfo_unexecuted_blocks=1 00:03:11.552 00:03:11.552 ' 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:11.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:11.552 --rc genhtml_branch_coverage=1 00:03:11.552 --rc genhtml_function_coverage=1 00:03:11.552 --rc genhtml_legend=1 00:03:11.552 --rc geninfo_all_blocks=1 00:03:11.552 --rc geninfo_unexecuted_blocks=1 00:03:11.552 00:03:11.552 ' 00:03:11.552 17:24:19 env -- env/env.sh@10 -- # run_test env_memory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:11.552 17:24:19 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:11.552 17:24:19 env -- common/autotest_common.sh@10 -- # set +x 00:03:11.813 ************************************ 00:03:11.813 START TEST env_memory 00:03:11.813 ************************************ 00:03:11.813 17:24:19 env.env_memory -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:03:11.813 00:03:11.813 00:03:11.813 CUnit - A unit testing framework for C - Version 2.1-3 00:03:11.813 http://cunit.sourceforge.net/ 00:03:11.813 00:03:11.813 00:03:11.813 Suite: memory 00:03:11.813 Test: alloc and free memory map ...[2024-10-17 17:24:19.526654] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:03:11.813 passed 00:03:11.813 Test: mem map translation ...[2024-10-17 17:24:19.554913] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:03:11.813 [2024-10-17 17:24:19.554947] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:03:11.813 [2024-10-17 17:24:19.554995] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:03:11.813 [2024-10-17 17:24:19.555003] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:03:11.813 passed 00:03:11.813 Test: mem map registration ...[2024-10-17 17:24:19.615125] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:03:11.813 [2024-10-17 17:24:19.615150] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:03:11.813 passed 00:03:11.813 Test: mem map adjacent registrations ...passed 00:03:11.813 00:03:11.813 Run Summary: Type Total Ran Passed Failed Inactive 00:03:11.813 suites 1 1 n/a 0 0 00:03:11.813 tests 4 4 4 0 0 00:03:11.813 asserts 152 152 152 0 n/a 00:03:11.813 00:03:11.813 Elapsed time = 0.203 seconds 00:03:11.813 00:03:11.813 real 0m0.218s 00:03:11.813 user 0m0.206s 00:03:11.813 sys 0m0.011s 00:03:11.813 17:24:19 env.env_memory -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:11.813 17:24:19 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:03:11.813 ************************************ 00:03:11.813 END TEST env_memory 00:03:11.813 ************************************ 00:03:11.813 17:24:19 env -- env/env.sh@11 -- # run_test env_vtophys /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:03:11.813 17:24:19 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:11.813 17:24:19 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:11.813 17:24:19 env -- common/autotest_common.sh@10 -- # set +x 00:03:12.075 ************************************ 00:03:12.075 START TEST env_vtophys 00:03:12.075 ************************************ 00:03:12.075 17:24:19 env.env_vtophys -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:03:12.075 EAL: lib.eal log level changed from notice to debug 00:03:12.075 EAL: Detected lcore 0 as core 0 on socket 0 00:03:12.075 EAL: Detected lcore 1 as core 1 on socket 0 00:03:12.075 EAL: Detected lcore 2 as core 2 on socket 0 00:03:12.075 EAL: Detected lcore 3 as core 3 on socket 0 00:03:12.075 EAL: Detected lcore 4 as core 4 on socket 0 00:03:12.075 EAL: Detected lcore 5 as core 5 on socket 0 00:03:12.075 EAL: Detected lcore 6 as core 6 on socket 0 00:03:12.075 EAL: Detected lcore 7 as core 7 on socket 0 00:03:12.075 EAL: Detected lcore 8 as core 8 on socket 0 00:03:12.075 EAL: Detected lcore 9 as core 9 on socket 0 00:03:12.075 EAL: Detected lcore 10 as core 10 on socket 0 00:03:12.075 EAL: Detected lcore 11 as core 11 on socket 0 00:03:12.075 EAL: Detected lcore 12 as core 12 on socket 0 00:03:12.075 EAL: Detected lcore 13 as core 13 on socket 0 00:03:12.075 EAL: Detected lcore 14 as core 14 on socket 0 00:03:12.075 EAL: Detected lcore 15 as core 15 on socket 0 00:03:12.075 EAL: Detected lcore 16 as core 16 on socket 0 00:03:12.075 EAL: Detected lcore 17 as core 17 on socket 0 00:03:12.075 EAL: Detected lcore 18 as core 18 on socket 0 00:03:12.075 EAL: Detected lcore 19 as core 19 on socket 0 00:03:12.075 EAL: Detected lcore 20 as core 20 on socket 0 00:03:12.075 EAL: Detected lcore 21 as core 21 on socket 0 00:03:12.075 EAL: Detected lcore 22 as core 22 on socket 0 00:03:12.075 EAL: Detected lcore 23 as core 23 on socket 0 00:03:12.075 EAL: Detected lcore 24 as core 24 on socket 0 00:03:12.075 EAL: Detected lcore 25 as core 25 on socket 0 00:03:12.075 EAL: Detected lcore 26 as core 26 on socket 0 00:03:12.075 EAL: Detected lcore 27 as core 27 on socket 0 00:03:12.075 EAL: Detected lcore 28 as core 28 on socket 0 00:03:12.075 EAL: Detected lcore 29 as core 29 on socket 0 00:03:12.075 EAL: Detected lcore 30 as core 30 on socket 0 00:03:12.075 EAL: Detected lcore 31 as core 31 on socket 0 00:03:12.075 EAL: Detected lcore 32 as core 32 on socket 0 00:03:12.075 EAL: Detected lcore 33 as core 33 on socket 0 00:03:12.075 EAL: Detected lcore 34 as core 34 on socket 0 00:03:12.075 EAL: Detected lcore 35 as core 35 on socket 0 00:03:12.075 EAL: Detected lcore 36 as core 0 on socket 1 00:03:12.075 EAL: Detected lcore 37 as core 1 on socket 1 00:03:12.075 EAL: Detected lcore 38 as core 2 on socket 1 00:03:12.075 EAL: Detected lcore 39 as core 3 on socket 1 00:03:12.075 EAL: Detected lcore 40 as core 4 on socket 1 00:03:12.075 EAL: Detected lcore 41 as core 5 on socket 1 00:03:12.075 EAL: Detected lcore 42 as core 6 on socket 1 00:03:12.075 EAL: Detected lcore 43 as core 7 on socket 1 00:03:12.075 EAL: Detected lcore 44 as core 8 on socket 1 00:03:12.075 EAL: Detected lcore 45 as core 9 on socket 1 00:03:12.075 EAL: Detected lcore 46 as core 10 on socket 1 00:03:12.075 EAL: Detected lcore 47 as core 11 on socket 1 00:03:12.075 EAL: Detected lcore 48 as core 12 on socket 1 00:03:12.075 EAL: Detected lcore 49 as core 13 on socket 1 00:03:12.075 EAL: Detected lcore 50 as core 14 on socket 1 00:03:12.075 EAL: Detected lcore 51 as core 15 on socket 1 00:03:12.075 EAL: Detected lcore 52 as core 16 on socket 1 00:03:12.075 EAL: Detected lcore 53 as core 17 on socket 1 00:03:12.075 EAL: Detected lcore 54 as core 18 on socket 1 00:03:12.075 EAL: Detected lcore 55 as core 19 on socket 1 00:03:12.075 EAL: Detected lcore 56 as core 20 on socket 1 00:03:12.075 EAL: Detected lcore 57 as core 21 on socket 1 00:03:12.075 EAL: Detected lcore 58 as core 22 on socket 1 00:03:12.075 EAL: Detected lcore 59 as core 23 on socket 1 00:03:12.075 EAL: Detected lcore 60 as core 24 on socket 1 00:03:12.075 EAL: Detected lcore 61 as core 25 on socket 1 00:03:12.075 EAL: Detected lcore 62 as core 26 on socket 1 00:03:12.075 EAL: Detected lcore 63 as core 27 on socket 1 00:03:12.075 EAL: Detected lcore 64 as core 28 on socket 1 00:03:12.075 EAL: Detected lcore 65 as core 29 on socket 1 00:03:12.075 EAL: Detected lcore 66 as core 30 on socket 1 00:03:12.075 EAL: Detected lcore 67 as core 31 on socket 1 00:03:12.075 EAL: Detected lcore 68 as core 32 on socket 1 00:03:12.075 EAL: Detected lcore 69 as core 33 on socket 1 00:03:12.075 EAL: Detected lcore 70 as core 34 on socket 1 00:03:12.075 EAL: Detected lcore 71 as core 35 on socket 1 00:03:12.075 EAL: Detected lcore 72 as core 0 on socket 0 00:03:12.075 EAL: Detected lcore 73 as core 1 on socket 0 00:03:12.075 EAL: Detected lcore 74 as core 2 on socket 0 00:03:12.075 EAL: Detected lcore 75 as core 3 on socket 0 00:03:12.075 EAL: Detected lcore 76 as core 4 on socket 0 00:03:12.075 EAL: Detected lcore 77 as core 5 on socket 0 00:03:12.075 EAL: Detected lcore 78 as core 6 on socket 0 00:03:12.075 EAL: Detected lcore 79 as core 7 on socket 0 00:03:12.075 EAL: Detected lcore 80 as core 8 on socket 0 00:03:12.075 EAL: Detected lcore 81 as core 9 on socket 0 00:03:12.075 EAL: Detected lcore 82 as core 10 on socket 0 00:03:12.075 EAL: Detected lcore 83 as core 11 on socket 0 00:03:12.075 EAL: Detected lcore 84 as core 12 on socket 0 00:03:12.075 EAL: Detected lcore 85 as core 13 on socket 0 00:03:12.075 EAL: Detected lcore 86 as core 14 on socket 0 00:03:12.075 EAL: Detected lcore 87 as core 15 on socket 0 00:03:12.075 EAL: Detected lcore 88 as core 16 on socket 0 00:03:12.075 EAL: Detected lcore 89 as core 17 on socket 0 00:03:12.075 EAL: Detected lcore 90 as core 18 on socket 0 00:03:12.075 EAL: Detected lcore 91 as core 19 on socket 0 00:03:12.075 EAL: Detected lcore 92 as core 20 on socket 0 00:03:12.075 EAL: Detected lcore 93 as core 21 on socket 0 00:03:12.075 EAL: Detected lcore 94 as core 22 on socket 0 00:03:12.075 EAL: Detected lcore 95 as core 23 on socket 0 00:03:12.075 EAL: Detected lcore 96 as core 24 on socket 0 00:03:12.075 EAL: Detected lcore 97 as core 25 on socket 0 00:03:12.075 EAL: Detected lcore 98 as core 26 on socket 0 00:03:12.075 EAL: Detected lcore 99 as core 27 on socket 0 00:03:12.075 EAL: Detected lcore 100 as core 28 on socket 0 00:03:12.075 EAL: Detected lcore 101 as core 29 on socket 0 00:03:12.075 EAL: Detected lcore 102 as core 30 on socket 0 00:03:12.075 EAL: Detected lcore 103 as core 31 on socket 0 00:03:12.075 EAL: Detected lcore 104 as core 32 on socket 0 00:03:12.075 EAL: Detected lcore 105 as core 33 on socket 0 00:03:12.075 EAL: Detected lcore 106 as core 34 on socket 0 00:03:12.075 EAL: Detected lcore 107 as core 35 on socket 0 00:03:12.075 EAL: Detected lcore 108 as core 0 on socket 1 00:03:12.075 EAL: Detected lcore 109 as core 1 on socket 1 00:03:12.075 EAL: Detected lcore 110 as core 2 on socket 1 00:03:12.075 EAL: Detected lcore 111 as core 3 on socket 1 00:03:12.075 EAL: Detected lcore 112 as core 4 on socket 1 00:03:12.075 EAL: Detected lcore 113 as core 5 on socket 1 00:03:12.075 EAL: Detected lcore 114 as core 6 on socket 1 00:03:12.075 EAL: Detected lcore 115 as core 7 on socket 1 00:03:12.075 EAL: Detected lcore 116 as core 8 on socket 1 00:03:12.075 EAL: Detected lcore 117 as core 9 on socket 1 00:03:12.075 EAL: Detected lcore 118 as core 10 on socket 1 00:03:12.075 EAL: Detected lcore 119 as core 11 on socket 1 00:03:12.075 EAL: Detected lcore 120 as core 12 on socket 1 00:03:12.075 EAL: Detected lcore 121 as core 13 on socket 1 00:03:12.075 EAL: Detected lcore 122 as core 14 on socket 1 00:03:12.075 EAL: Detected lcore 123 as core 15 on socket 1 00:03:12.075 EAL: Detected lcore 124 as core 16 on socket 1 00:03:12.075 EAL: Detected lcore 125 as core 17 on socket 1 00:03:12.075 EAL: Detected lcore 126 as core 18 on socket 1 00:03:12.075 EAL: Detected lcore 127 as core 19 on socket 1 00:03:12.075 EAL: Skipped lcore 128 as core 20 on socket 1 00:03:12.075 EAL: Skipped lcore 129 as core 21 on socket 1 00:03:12.075 EAL: Skipped lcore 130 as core 22 on socket 1 00:03:12.075 EAL: Skipped lcore 131 as core 23 on socket 1 00:03:12.075 EAL: Skipped lcore 132 as core 24 on socket 1 00:03:12.075 EAL: Skipped lcore 133 as core 25 on socket 1 00:03:12.075 EAL: Skipped lcore 134 as core 26 on socket 1 00:03:12.075 EAL: Skipped lcore 135 as core 27 on socket 1 00:03:12.075 EAL: Skipped lcore 136 as core 28 on socket 1 00:03:12.075 EAL: Skipped lcore 137 as core 29 on socket 1 00:03:12.075 EAL: Skipped lcore 138 as core 30 on socket 1 00:03:12.075 EAL: Skipped lcore 139 as core 31 on socket 1 00:03:12.075 EAL: Skipped lcore 140 as core 32 on socket 1 00:03:12.075 EAL: Skipped lcore 141 as core 33 on socket 1 00:03:12.075 EAL: Skipped lcore 142 as core 34 on socket 1 00:03:12.075 EAL: Skipped lcore 143 as core 35 on socket 1 00:03:12.075 EAL: Maximum logical cores by configuration: 128 00:03:12.075 EAL: Detected CPU lcores: 128 00:03:12.075 EAL: Detected NUMA nodes: 2 00:03:12.075 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:03:12.075 EAL: Detected shared linkage of DPDK 00:03:12.075 EAL: No shared files mode enabled, IPC will be disabled 00:03:12.075 EAL: Bus pci wants IOVA as 'DC' 00:03:12.075 EAL: Buses did not request a specific IOVA mode. 00:03:12.075 EAL: IOMMU is available, selecting IOVA as VA mode. 00:03:12.075 EAL: Selected IOVA mode 'VA' 00:03:12.075 EAL: Probing VFIO support... 00:03:12.075 EAL: IOMMU type 1 (Type 1) is supported 00:03:12.075 EAL: IOMMU type 7 (sPAPR) is not supported 00:03:12.075 EAL: IOMMU type 8 (No-IOMMU) is not supported 00:03:12.075 EAL: VFIO support initialized 00:03:12.075 EAL: Ask a virtual area of 0x2e000 bytes 00:03:12.075 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:03:12.075 EAL: Setting up physically contiguous memory... 00:03:12.075 EAL: Setting maximum number of open files to 524288 00:03:12.075 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:03:12.075 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 00:03:12.075 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:03:12.075 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.075 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:03:12.075 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:12.075 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.075 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:03:12.075 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:03:12.075 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.075 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:03:12.075 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:12.075 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.075 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:03:12.075 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:03:12.075 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.075 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:03:12.075 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:12.075 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.075 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:03:12.075 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:03:12.075 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.075 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:03:12.076 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:03:12.076 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.076 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:03:12.076 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:03:12.076 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 00:03:12.076 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.076 EAL: Virtual area found at 0x201000800000 (size = 0x61000) 00:03:12.076 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:12.076 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.076 EAL: Virtual area found at 0x201000a00000 (size = 0x400000000) 00:03:12.076 EAL: VA reserved for memseg list at 0x201000a00000, size 400000000 00:03:12.076 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.076 EAL: Virtual area found at 0x201400a00000 (size = 0x61000) 00:03:12.076 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:12.076 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.076 EAL: Virtual area found at 0x201400c00000 (size = 0x400000000) 00:03:12.076 EAL: VA reserved for memseg list at 0x201400c00000, size 400000000 00:03:12.076 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.076 EAL: Virtual area found at 0x201800c00000 (size = 0x61000) 00:03:12.076 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:12.076 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.076 EAL: Virtual area found at 0x201800e00000 (size = 0x400000000) 00:03:12.076 EAL: VA reserved for memseg list at 0x201800e00000, size 400000000 00:03:12.076 EAL: Ask a virtual area of 0x61000 bytes 00:03:12.076 EAL: Virtual area found at 0x201c00e00000 (size = 0x61000) 00:03:12.076 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:03:12.076 EAL: Ask a virtual area of 0x400000000 bytes 00:03:12.076 EAL: Virtual area found at 0x201c01000000 (size = 0x400000000) 00:03:12.076 EAL: VA reserved for memseg list at 0x201c01000000, size 400000000 00:03:12.076 EAL: Hugepages will be freed exactly as allocated. 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: TSC frequency is ~2400000 KHz 00:03:12.076 EAL: Main lcore 0 is ready (tid=7f199d074a00;cpuset=[0]) 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 0 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 2MB 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: No PCI address specified using 'addr=' in: bus=pci 00:03:12.076 EAL: Mem event callback 'spdk:(nil)' registered 00:03:12.076 00:03:12.076 00:03:12.076 CUnit - A unit testing framework for C - Version 2.1-3 00:03:12.076 http://cunit.sourceforge.net/ 00:03:12.076 00:03:12.076 00:03:12.076 Suite: components_suite 00:03:12.076 Test: vtophys_malloc_test ...passed 00:03:12.076 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 4 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 4MB 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was shrunk by 4MB 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 4 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 6MB 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was shrunk by 6MB 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 4 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 10MB 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was shrunk by 10MB 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 4 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 18MB 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was shrunk by 18MB 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 4 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 34MB 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was shrunk by 34MB 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 4 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 66MB 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was shrunk by 66MB 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.076 EAL: Restoring previous memory policy: 4 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was expanded by 130MB 00:03:12.076 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.076 EAL: request: mp_malloc_sync 00:03:12.076 EAL: No shared files mode enabled, IPC is disabled 00:03:12.076 EAL: Heap on socket 0 was shrunk by 130MB 00:03:12.076 EAL: Trying to obtain current memory policy. 00:03:12.076 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.336 EAL: Restoring previous memory policy: 4 00:03:12.336 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.336 EAL: request: mp_malloc_sync 00:03:12.336 EAL: No shared files mode enabled, IPC is disabled 00:03:12.336 EAL: Heap on socket 0 was expanded by 258MB 00:03:12.336 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.336 EAL: request: mp_malloc_sync 00:03:12.336 EAL: No shared files mode enabled, IPC is disabled 00:03:12.336 EAL: Heap on socket 0 was shrunk by 258MB 00:03:12.336 EAL: Trying to obtain current memory policy. 00:03:12.336 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.336 EAL: Restoring previous memory policy: 4 00:03:12.336 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.336 EAL: request: mp_malloc_sync 00:03:12.336 EAL: No shared files mode enabled, IPC is disabled 00:03:12.336 EAL: Heap on socket 0 was expanded by 514MB 00:03:12.336 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.336 EAL: request: mp_malloc_sync 00:03:12.336 EAL: No shared files mode enabled, IPC is disabled 00:03:12.336 EAL: Heap on socket 0 was shrunk by 514MB 00:03:12.336 EAL: Trying to obtain current memory policy. 00:03:12.336 EAL: Setting policy MPOL_PREFERRED for socket 0 00:03:12.596 EAL: Restoring previous memory policy: 4 00:03:12.596 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.596 EAL: request: mp_malloc_sync 00:03:12.596 EAL: No shared files mode enabled, IPC is disabled 00:03:12.596 EAL: Heap on socket 0 was expanded by 1026MB 00:03:12.596 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.856 EAL: request: mp_malloc_sync 00:03:12.856 EAL: No shared files mode enabled, IPC is disabled 00:03:12.856 EAL: Heap on socket 0 was shrunk by 1026MB 00:03:12.856 passed 00:03:12.856 00:03:12.856 Run Summary: Type Total Ran Passed Failed Inactive 00:03:12.856 suites 1 1 n/a 0 0 00:03:12.856 tests 2 2 2 0 0 00:03:12.856 asserts 497 497 497 0 n/a 00:03:12.856 00:03:12.856 Elapsed time = 0.698 seconds 00:03:12.856 EAL: Calling mem event callback 'spdk:(nil)' 00:03:12.856 EAL: request: mp_malloc_sync 00:03:12.856 EAL: No shared files mode enabled, IPC is disabled 00:03:12.856 EAL: Heap on socket 0 was shrunk by 2MB 00:03:12.856 EAL: No shared files mode enabled, IPC is disabled 00:03:12.856 EAL: No shared files mode enabled, IPC is disabled 00:03:12.856 EAL: No shared files mode enabled, IPC is disabled 00:03:12.856 00:03:12.856 real 0m0.846s 00:03:12.856 user 0m0.440s 00:03:12.856 sys 0m0.373s 00:03:12.856 17:24:20 env.env_vtophys -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:12.856 17:24:20 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:03:12.856 ************************************ 00:03:12.856 END TEST env_vtophys 00:03:12.856 ************************************ 00:03:12.856 17:24:20 env -- env/env.sh@12 -- # run_test env_pci /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:03:12.856 17:24:20 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:12.856 17:24:20 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:12.856 17:24:20 env -- common/autotest_common.sh@10 -- # set +x 00:03:12.856 ************************************ 00:03:12.856 START TEST env_pci 00:03:12.856 ************************************ 00:03:12.856 17:24:20 env.env_pci -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:03:12.856 00:03:12.856 00:03:12.856 CUnit - A unit testing framework for C - Version 2.1-3 00:03:12.856 http://cunit.sourceforge.net/ 00:03:12.856 00:03:12.856 00:03:12.856 Suite: pci 00:03:12.856 Test: pci_hook ...[2024-10-17 17:24:20.713897] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/pci.c:1049:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 3997008 has claimed it 00:03:12.856 EAL: Cannot find device (10000:00:01.0) 00:03:12.856 EAL: Failed to attach device on primary process 00:03:12.856 passed 00:03:12.856 00:03:12.856 Run Summary: Type Total Ran Passed Failed Inactive 00:03:12.856 suites 1 1 n/a 0 0 00:03:12.856 tests 1 1 1 0 0 00:03:12.856 asserts 25 25 25 0 n/a 00:03:12.856 00:03:12.856 Elapsed time = 0.030 seconds 00:03:12.856 00:03:12.856 real 0m0.051s 00:03:12.856 user 0m0.017s 00:03:12.856 sys 0m0.033s 00:03:12.856 17:24:20 env.env_pci -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:12.856 17:24:20 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:03:12.856 ************************************ 00:03:12.856 END TEST env_pci 00:03:12.856 ************************************ 00:03:13.116 17:24:20 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:03:13.116 17:24:20 env -- env/env.sh@15 -- # uname 00:03:13.116 17:24:20 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:03:13.116 17:24:20 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:03:13.116 17:24:20 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:03:13.116 17:24:20 env -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:03:13.116 17:24:20 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:13.116 17:24:20 env -- common/autotest_common.sh@10 -- # set +x 00:03:13.116 ************************************ 00:03:13.116 START TEST env_dpdk_post_init 00:03:13.116 ************************************ 00:03:13.116 17:24:20 env.env_dpdk_post_init -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:03:13.116 EAL: Detected CPU lcores: 128 00:03:13.116 EAL: Detected NUMA nodes: 2 00:03:13.116 EAL: Detected shared linkage of DPDK 00:03:13.116 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:03:13.116 EAL: Selected IOVA mode 'VA' 00:03:13.116 EAL: VFIO support initialized 00:03:13.116 TELEMETRY: No legacy callbacks, legacy socket not created 00:03:13.116 EAL: Using IOMMU type 1 (Type 1) 00:03:13.376 EAL: Ignore mapping IO port bar(1) 00:03:13.376 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.0 (socket 0) 00:03:13.637 EAL: Ignore mapping IO port bar(1) 00:03:13.637 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.1 (socket 0) 00:03:13.637 EAL: Ignore mapping IO port bar(1) 00:03:13.898 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.2 (socket 0) 00:03:13.898 EAL: Ignore mapping IO port bar(1) 00:03:14.159 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.3 (socket 0) 00:03:14.159 EAL: Ignore mapping IO port bar(1) 00:03:14.419 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.4 (socket 0) 00:03:14.419 EAL: Ignore mapping IO port bar(1) 00:03:14.419 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.5 (socket 0) 00:03:14.679 EAL: Ignore mapping IO port bar(1) 00:03:14.679 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.6 (socket 0) 00:03:14.940 EAL: Ignore mapping IO port bar(1) 00:03:14.940 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.7 (socket 0) 00:03:15.200 EAL: Probe PCI driver: spdk_nvme (144d:a80a) device: 0000:65:00.0 (socket 0) 00:03:15.200 EAL: Ignore mapping IO port bar(1) 00:03:15.460 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.0 (socket 1) 00:03:15.460 EAL: Ignore mapping IO port bar(1) 00:03:15.721 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.1 (socket 1) 00:03:15.721 EAL: Ignore mapping IO port bar(1) 00:03:15.989 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.2 (socket 1) 00:03:15.989 EAL: Ignore mapping IO port bar(1) 00:03:15.989 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.3 (socket 1) 00:03:16.265 EAL: Ignore mapping IO port bar(1) 00:03:16.265 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.4 (socket 1) 00:03:16.575 EAL: Ignore mapping IO port bar(1) 00:03:16.575 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.5 (socket 1) 00:03:16.575 EAL: Ignore mapping IO port bar(1) 00:03:16.836 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.6 (socket 1) 00:03:16.836 EAL: Ignore mapping IO port bar(1) 00:03:17.098 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.7 (socket 1) 00:03:17.098 EAL: Releasing PCI mapped resource for 0000:65:00.0 00:03:17.098 EAL: Calling pci_unmap_resource for 0000:65:00.0 at 0x202001020000 00:03:17.098 Starting DPDK initialization... 00:03:17.098 Starting SPDK post initialization... 00:03:17.098 SPDK NVMe probe 00:03:17.098 Attaching to 0000:65:00.0 00:03:17.098 Attached to 0000:65:00.0 00:03:17.098 Cleaning up... 00:03:19.012 00:03:19.012 real 0m5.740s 00:03:19.012 user 0m0.093s 00:03:19.012 sys 0m0.204s 00:03:19.012 17:24:26 env.env_dpdk_post_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:19.012 17:24:26 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:03:19.012 ************************************ 00:03:19.012 END TEST env_dpdk_post_init 00:03:19.012 ************************************ 00:03:19.012 17:24:26 env -- env/env.sh@26 -- # uname 00:03:19.012 17:24:26 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:03:19.012 17:24:26 env -- env/env.sh@29 -- # run_test env_mem_callbacks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:03:19.012 17:24:26 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:19.012 17:24:26 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:19.012 17:24:26 env -- common/autotest_common.sh@10 -- # set +x 00:03:19.012 ************************************ 00:03:19.012 START TEST env_mem_callbacks 00:03:19.012 ************************************ 00:03:19.012 17:24:26 env.env_mem_callbacks -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:03:19.012 EAL: Detected CPU lcores: 128 00:03:19.012 EAL: Detected NUMA nodes: 2 00:03:19.012 EAL: Detected shared linkage of DPDK 00:03:19.012 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:03:19.012 EAL: Selected IOVA mode 'VA' 00:03:19.012 EAL: VFIO support initialized 00:03:19.012 TELEMETRY: No legacy callbacks, legacy socket not created 00:03:19.012 00:03:19.012 00:03:19.012 CUnit - A unit testing framework for C - Version 2.1-3 00:03:19.012 http://cunit.sourceforge.net/ 00:03:19.012 00:03:19.012 00:03:19.012 Suite: memory 00:03:19.012 Test: test ... 00:03:19.012 register 0x200000200000 2097152 00:03:19.012 malloc 3145728 00:03:19.012 register 0x200000400000 4194304 00:03:19.012 buf 0x200000500000 len 3145728 PASSED 00:03:19.012 malloc 64 00:03:19.012 buf 0x2000004fff40 len 64 PASSED 00:03:19.012 malloc 4194304 00:03:19.012 register 0x200000800000 6291456 00:03:19.012 buf 0x200000a00000 len 4194304 PASSED 00:03:19.012 free 0x200000500000 3145728 00:03:19.012 free 0x2000004fff40 64 00:03:19.012 unregister 0x200000400000 4194304 PASSED 00:03:19.012 free 0x200000a00000 4194304 00:03:19.012 unregister 0x200000800000 6291456 PASSED 00:03:19.012 malloc 8388608 00:03:19.012 register 0x200000400000 10485760 00:03:19.012 buf 0x200000600000 len 8388608 PASSED 00:03:19.012 free 0x200000600000 8388608 00:03:19.012 unregister 0x200000400000 10485760 PASSED 00:03:19.012 passed 00:03:19.012 00:03:19.012 Run Summary: Type Total Ran Passed Failed Inactive 00:03:19.012 suites 1 1 n/a 0 0 00:03:19.012 tests 1 1 1 0 0 00:03:19.012 asserts 15 15 15 0 n/a 00:03:19.012 00:03:19.012 Elapsed time = 0.010 seconds 00:03:19.012 00:03:19.012 real 0m0.072s 00:03:19.012 user 0m0.022s 00:03:19.012 sys 0m0.050s 00:03:19.012 17:24:26 env.env_mem_callbacks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:19.012 17:24:26 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:03:19.012 ************************************ 00:03:19.012 END TEST env_mem_callbacks 00:03:19.012 ************************************ 00:03:19.012 00:03:19.012 real 0m7.548s 00:03:19.012 user 0m1.043s 00:03:19.012 sys 0m1.064s 00:03:19.012 17:24:26 env -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:19.012 17:24:26 env -- common/autotest_common.sh@10 -- # set +x 00:03:19.012 ************************************ 00:03:19.012 END TEST env 00:03:19.012 ************************************ 00:03:19.012 17:24:26 -- spdk/autotest.sh@156 -- # run_test rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:03:19.012 17:24:26 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:19.012 17:24:26 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:19.012 17:24:26 -- common/autotest_common.sh@10 -- # set +x 00:03:19.012 ************************************ 00:03:19.012 START TEST rpc 00:03:19.012 ************************************ 00:03:19.012 17:24:26 rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:03:19.275 * Looking for test storage... 00:03:19.275 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:03:19.275 17:24:26 rpc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:19.275 17:24:26 rpc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:19.275 17:24:26 rpc -- common/autotest_common.sh@1691 -- # lcov --version 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:19.275 17:24:27 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:19.275 17:24:27 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:03:19.275 17:24:27 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:03:19.275 17:24:27 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:03:19.275 17:24:27 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:03:19.275 17:24:27 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:03:19.275 17:24:27 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:19.275 17:24:27 rpc -- scripts/common.sh@344 -- # case "$op" in 00:03:19.275 17:24:27 rpc -- scripts/common.sh@345 -- # : 1 00:03:19.275 17:24:27 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:19.275 17:24:27 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:19.275 17:24:27 rpc -- scripts/common.sh@365 -- # decimal 1 00:03:19.275 17:24:27 rpc -- scripts/common.sh@353 -- # local d=1 00:03:19.275 17:24:27 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:19.275 17:24:27 rpc -- scripts/common.sh@355 -- # echo 1 00:03:19.275 17:24:27 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:03:19.275 17:24:27 rpc -- scripts/common.sh@366 -- # decimal 2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@353 -- # local d=2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:19.275 17:24:27 rpc -- scripts/common.sh@355 -- # echo 2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:03:19.275 17:24:27 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:19.275 17:24:27 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:19.275 17:24:27 rpc -- scripts/common.sh@368 -- # return 0 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:19.275 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:19.275 --rc genhtml_branch_coverage=1 00:03:19.275 --rc genhtml_function_coverage=1 00:03:19.275 --rc genhtml_legend=1 00:03:19.275 --rc geninfo_all_blocks=1 00:03:19.275 --rc geninfo_unexecuted_blocks=1 00:03:19.275 00:03:19.275 ' 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:19.275 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:19.275 --rc genhtml_branch_coverage=1 00:03:19.275 --rc genhtml_function_coverage=1 00:03:19.275 --rc genhtml_legend=1 00:03:19.275 --rc geninfo_all_blocks=1 00:03:19.275 --rc geninfo_unexecuted_blocks=1 00:03:19.275 00:03:19.275 ' 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:19.275 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:19.275 --rc genhtml_branch_coverage=1 00:03:19.275 --rc genhtml_function_coverage=1 00:03:19.275 --rc genhtml_legend=1 00:03:19.275 --rc geninfo_all_blocks=1 00:03:19.275 --rc geninfo_unexecuted_blocks=1 00:03:19.275 00:03:19.275 ' 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:19.275 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:19.275 --rc genhtml_branch_coverage=1 00:03:19.275 --rc genhtml_function_coverage=1 00:03:19.275 --rc genhtml_legend=1 00:03:19.275 --rc geninfo_all_blocks=1 00:03:19.275 --rc geninfo_unexecuted_blocks=1 00:03:19.275 00:03:19.275 ' 00:03:19.275 17:24:27 rpc -- rpc/rpc.sh@65 -- # spdk_pid=3998328 00:03:19.275 17:24:27 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:03:19.275 17:24:27 rpc -- rpc/rpc.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -e bdev 00:03:19.275 17:24:27 rpc -- rpc/rpc.sh@67 -- # waitforlisten 3998328 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@831 -- # '[' -z 3998328 ']' 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:03:19.275 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:19.275 17:24:27 rpc -- common/autotest_common.sh@10 -- # set +x 00:03:19.275 [2024-10-17 17:24:27.126769] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:19.275 [2024-10-17 17:24:27.126840] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3998328 ] 00:03:19.536 [2024-10-17 17:24:27.210351] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:19.536 [2024-10-17 17:24:27.262740] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:03:19.536 [2024-10-17 17:24:27.262797] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 3998328' to capture a snapshot of events at runtime. 00:03:19.536 [2024-10-17 17:24:27.262806] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:03:19.536 [2024-10-17 17:24:27.262814] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:03:19.536 [2024-10-17 17:24:27.262821] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid3998328 for offline analysis/debug. 00:03:19.536 [2024-10-17 17:24:27.263626] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:20.108 17:24:27 rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:20.108 17:24:27 rpc -- common/autotest_common.sh@864 -- # return 0 00:03:20.108 17:24:27 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:03:20.108 17:24:27 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:03:20.108 17:24:27 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:03:20.108 17:24:27 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:03:20.108 17:24:27 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:20.108 17:24:27 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:20.108 17:24:27 rpc -- common/autotest_common.sh@10 -- # set +x 00:03:20.108 ************************************ 00:03:20.108 START TEST rpc_integrity 00:03:20.108 ************************************ 00:03:20.108 17:24:27 rpc.rpc_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:03:20.108 17:24:27 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:03:20.108 17:24:27 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.108 17:24:27 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.108 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.108 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:03:20.108 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:03:20.370 { 00:03:20.370 "name": "Malloc0", 00:03:20.370 "aliases": [ 00:03:20.370 "7e26642c-6ecc-4273-b185-8f547cc394cd" 00:03:20.370 ], 00:03:20.370 "product_name": "Malloc disk", 00:03:20.370 "block_size": 512, 00:03:20.370 "num_blocks": 16384, 00:03:20.370 "uuid": "7e26642c-6ecc-4273-b185-8f547cc394cd", 00:03:20.370 "assigned_rate_limits": { 00:03:20.370 "rw_ios_per_sec": 0, 00:03:20.370 "rw_mbytes_per_sec": 0, 00:03:20.370 "r_mbytes_per_sec": 0, 00:03:20.370 "w_mbytes_per_sec": 0 00:03:20.370 }, 00:03:20.370 "claimed": false, 00:03:20.370 "zoned": false, 00:03:20.370 "supported_io_types": { 00:03:20.370 "read": true, 00:03:20.370 "write": true, 00:03:20.370 "unmap": true, 00:03:20.370 "flush": true, 00:03:20.370 "reset": true, 00:03:20.370 "nvme_admin": false, 00:03:20.370 "nvme_io": false, 00:03:20.370 "nvme_io_md": false, 00:03:20.370 "write_zeroes": true, 00:03:20.370 "zcopy": true, 00:03:20.370 "get_zone_info": false, 00:03:20.370 "zone_management": false, 00:03:20.370 "zone_append": false, 00:03:20.370 "compare": false, 00:03:20.370 "compare_and_write": false, 00:03:20.370 "abort": true, 00:03:20.370 "seek_hole": false, 00:03:20.370 "seek_data": false, 00:03:20.370 "copy": true, 00:03:20.370 "nvme_iov_md": false 00:03:20.370 }, 00:03:20.370 "memory_domains": [ 00:03:20.370 { 00:03:20.370 "dma_device_id": "system", 00:03:20.370 "dma_device_type": 1 00:03:20.370 }, 00:03:20.370 { 00:03:20.370 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:03:20.370 "dma_device_type": 2 00:03:20.370 } 00:03:20.370 ], 00:03:20.370 "driver_specific": {} 00:03:20.370 } 00:03:20.370 ]' 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.370 [2024-10-17 17:24:28.133398] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:03:20.370 [2024-10-17 17:24:28.133448] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:03:20.370 [2024-10-17 17:24:28.133465] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x15e7b00 00:03:20.370 [2024-10-17 17:24:28.133473] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:03:20.370 [2024-10-17 17:24:28.135070] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:03:20.370 [2024-10-17 17:24:28.135108] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:03:20.370 Passthru0 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.370 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.370 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:03:20.370 { 00:03:20.370 "name": "Malloc0", 00:03:20.370 "aliases": [ 00:03:20.370 "7e26642c-6ecc-4273-b185-8f547cc394cd" 00:03:20.370 ], 00:03:20.370 "product_name": "Malloc disk", 00:03:20.370 "block_size": 512, 00:03:20.370 "num_blocks": 16384, 00:03:20.370 "uuid": "7e26642c-6ecc-4273-b185-8f547cc394cd", 00:03:20.370 "assigned_rate_limits": { 00:03:20.370 "rw_ios_per_sec": 0, 00:03:20.370 "rw_mbytes_per_sec": 0, 00:03:20.370 "r_mbytes_per_sec": 0, 00:03:20.370 "w_mbytes_per_sec": 0 00:03:20.370 }, 00:03:20.370 "claimed": true, 00:03:20.370 "claim_type": "exclusive_write", 00:03:20.370 "zoned": false, 00:03:20.370 "supported_io_types": { 00:03:20.370 "read": true, 00:03:20.370 "write": true, 00:03:20.370 "unmap": true, 00:03:20.370 "flush": true, 00:03:20.370 "reset": true, 00:03:20.370 "nvme_admin": false, 00:03:20.370 "nvme_io": false, 00:03:20.370 "nvme_io_md": false, 00:03:20.370 "write_zeroes": true, 00:03:20.370 "zcopy": true, 00:03:20.370 "get_zone_info": false, 00:03:20.370 "zone_management": false, 00:03:20.370 "zone_append": false, 00:03:20.370 "compare": false, 00:03:20.370 "compare_and_write": false, 00:03:20.370 "abort": true, 00:03:20.370 "seek_hole": false, 00:03:20.370 "seek_data": false, 00:03:20.370 "copy": true, 00:03:20.370 "nvme_iov_md": false 00:03:20.370 }, 00:03:20.370 "memory_domains": [ 00:03:20.370 { 00:03:20.370 "dma_device_id": "system", 00:03:20.370 "dma_device_type": 1 00:03:20.370 }, 00:03:20.370 { 00:03:20.370 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:03:20.370 "dma_device_type": 2 00:03:20.370 } 00:03:20.370 ], 00:03:20.370 "driver_specific": {} 00:03:20.370 }, 00:03:20.370 { 00:03:20.370 "name": "Passthru0", 00:03:20.370 "aliases": [ 00:03:20.370 "51bfebc6-6f7b-56b9-9ae5-733aedc91cfb" 00:03:20.370 ], 00:03:20.370 "product_name": "passthru", 00:03:20.370 "block_size": 512, 00:03:20.370 "num_blocks": 16384, 00:03:20.370 "uuid": "51bfebc6-6f7b-56b9-9ae5-733aedc91cfb", 00:03:20.370 "assigned_rate_limits": { 00:03:20.370 "rw_ios_per_sec": 0, 00:03:20.370 "rw_mbytes_per_sec": 0, 00:03:20.370 "r_mbytes_per_sec": 0, 00:03:20.370 "w_mbytes_per_sec": 0 00:03:20.370 }, 00:03:20.370 "claimed": false, 00:03:20.370 "zoned": false, 00:03:20.370 "supported_io_types": { 00:03:20.370 "read": true, 00:03:20.370 "write": true, 00:03:20.370 "unmap": true, 00:03:20.370 "flush": true, 00:03:20.370 "reset": true, 00:03:20.370 "nvme_admin": false, 00:03:20.370 "nvme_io": false, 00:03:20.370 "nvme_io_md": false, 00:03:20.370 "write_zeroes": true, 00:03:20.370 "zcopy": true, 00:03:20.370 "get_zone_info": false, 00:03:20.370 "zone_management": false, 00:03:20.370 "zone_append": false, 00:03:20.370 "compare": false, 00:03:20.370 "compare_and_write": false, 00:03:20.370 "abort": true, 00:03:20.370 "seek_hole": false, 00:03:20.370 "seek_data": false, 00:03:20.370 "copy": true, 00:03:20.370 "nvme_iov_md": false 00:03:20.370 }, 00:03:20.370 "memory_domains": [ 00:03:20.370 { 00:03:20.370 "dma_device_id": "system", 00:03:20.370 "dma_device_type": 1 00:03:20.370 }, 00:03:20.370 { 00:03:20.370 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:03:20.370 "dma_device_type": 2 00:03:20.370 } 00:03:20.370 ], 00:03:20.370 "driver_specific": { 00:03:20.370 "passthru": { 00:03:20.370 "name": "Passthru0", 00:03:20.371 "base_bdev_name": "Malloc0" 00:03:20.371 } 00:03:20.371 } 00:03:20.371 } 00:03:20.371 ]' 00:03:20.371 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:03:20.371 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:03:20.371 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.371 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.371 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.371 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.371 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:03:20.371 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:03:20.632 17:24:28 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:03:20.632 00:03:20.632 real 0m0.300s 00:03:20.632 user 0m0.183s 00:03:20.632 sys 0m0.048s 00:03:20.632 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:20.632 17:24:28 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:20.632 ************************************ 00:03:20.632 END TEST rpc_integrity 00:03:20.632 ************************************ 00:03:20.632 17:24:28 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:03:20.632 17:24:28 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:20.632 17:24:28 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:20.632 17:24:28 rpc -- common/autotest_common.sh@10 -- # set +x 00:03:20.632 ************************************ 00:03:20.632 START TEST rpc_plugins 00:03:20.632 ************************************ 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@1125 -- # rpc_plugins 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:03:20.632 { 00:03:20.632 "name": "Malloc1", 00:03:20.632 "aliases": [ 00:03:20.632 "7b712e0f-fb37-4a2c-9a8b-5c307ac3ab3a" 00:03:20.632 ], 00:03:20.632 "product_name": "Malloc disk", 00:03:20.632 "block_size": 4096, 00:03:20.632 "num_blocks": 256, 00:03:20.632 "uuid": "7b712e0f-fb37-4a2c-9a8b-5c307ac3ab3a", 00:03:20.632 "assigned_rate_limits": { 00:03:20.632 "rw_ios_per_sec": 0, 00:03:20.632 "rw_mbytes_per_sec": 0, 00:03:20.632 "r_mbytes_per_sec": 0, 00:03:20.632 "w_mbytes_per_sec": 0 00:03:20.632 }, 00:03:20.632 "claimed": false, 00:03:20.632 "zoned": false, 00:03:20.632 "supported_io_types": { 00:03:20.632 "read": true, 00:03:20.632 "write": true, 00:03:20.632 "unmap": true, 00:03:20.632 "flush": true, 00:03:20.632 "reset": true, 00:03:20.632 "nvme_admin": false, 00:03:20.632 "nvme_io": false, 00:03:20.632 "nvme_io_md": false, 00:03:20.632 "write_zeroes": true, 00:03:20.632 "zcopy": true, 00:03:20.632 "get_zone_info": false, 00:03:20.632 "zone_management": false, 00:03:20.632 "zone_append": false, 00:03:20.632 "compare": false, 00:03:20.632 "compare_and_write": false, 00:03:20.632 "abort": true, 00:03:20.632 "seek_hole": false, 00:03:20.632 "seek_data": false, 00:03:20.632 "copy": true, 00:03:20.632 "nvme_iov_md": false 00:03:20.632 }, 00:03:20.632 "memory_domains": [ 00:03:20.632 { 00:03:20.632 "dma_device_id": "system", 00:03:20.632 "dma_device_type": 1 00:03:20.632 }, 00:03:20.632 { 00:03:20.632 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:03:20.632 "dma_device_type": 2 00:03:20.632 } 00:03:20.632 ], 00:03:20.632 "driver_specific": {} 00:03:20.632 } 00:03:20.632 ]' 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:03:20.632 17:24:28 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:03:20.632 00:03:20.632 real 0m0.154s 00:03:20.632 user 0m0.094s 00:03:20.632 sys 0m0.023s 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:20.632 17:24:28 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:03:20.632 ************************************ 00:03:20.632 END TEST rpc_plugins 00:03:20.632 ************************************ 00:03:20.893 17:24:28 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:03:20.893 17:24:28 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:20.893 17:24:28 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:20.893 17:24:28 rpc -- common/autotest_common.sh@10 -- # set +x 00:03:20.893 ************************************ 00:03:20.893 START TEST rpc_trace_cmd_test 00:03:20.893 ************************************ 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1125 -- # rpc_trace_cmd_test 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:03:20.893 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid3998328", 00:03:20.893 "tpoint_group_mask": "0x8", 00:03:20.893 "iscsi_conn": { 00:03:20.893 "mask": "0x2", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "scsi": { 00:03:20.893 "mask": "0x4", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "bdev": { 00:03:20.893 "mask": "0x8", 00:03:20.893 "tpoint_mask": "0xffffffffffffffff" 00:03:20.893 }, 00:03:20.893 "nvmf_rdma": { 00:03:20.893 "mask": "0x10", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "nvmf_tcp": { 00:03:20.893 "mask": "0x20", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "ftl": { 00:03:20.893 "mask": "0x40", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "blobfs": { 00:03:20.893 "mask": "0x80", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "dsa": { 00:03:20.893 "mask": "0x200", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "thread": { 00:03:20.893 "mask": "0x400", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "nvme_pcie": { 00:03:20.893 "mask": "0x800", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "iaa": { 00:03:20.893 "mask": "0x1000", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "nvme_tcp": { 00:03:20.893 "mask": "0x2000", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "bdev_nvme": { 00:03:20.893 "mask": "0x4000", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "sock": { 00:03:20.893 "mask": "0x8000", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "blob": { 00:03:20.893 "mask": "0x10000", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "bdev_raid": { 00:03:20.893 "mask": "0x20000", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 }, 00:03:20.893 "scheduler": { 00:03:20.893 "mask": "0x40000", 00:03:20.893 "tpoint_mask": "0x0" 00:03:20.893 } 00:03:20.893 }' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:03:20.893 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:03:21.155 17:24:28 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:03:21.155 00:03:21.155 real 0m0.233s 00:03:21.155 user 0m0.192s 00:03:21.155 sys 0m0.034s 00:03:21.155 17:24:28 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:21.155 17:24:28 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:03:21.155 ************************************ 00:03:21.155 END TEST rpc_trace_cmd_test 00:03:21.155 ************************************ 00:03:21.155 17:24:28 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:03:21.155 17:24:28 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:03:21.155 17:24:28 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:03:21.155 17:24:28 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:21.155 17:24:28 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:21.155 17:24:28 rpc -- common/autotest_common.sh@10 -- # set +x 00:03:21.155 ************************************ 00:03:21.155 START TEST rpc_daemon_integrity 00:03:21.155 ************************************ 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.155 17:24:28 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:03:21.155 { 00:03:21.155 "name": "Malloc2", 00:03:21.155 "aliases": [ 00:03:21.155 "309ce199-ea2c-4cce-985b-144b698da108" 00:03:21.155 ], 00:03:21.155 "product_name": "Malloc disk", 00:03:21.155 "block_size": 512, 00:03:21.155 "num_blocks": 16384, 00:03:21.155 "uuid": "309ce199-ea2c-4cce-985b-144b698da108", 00:03:21.155 "assigned_rate_limits": { 00:03:21.155 "rw_ios_per_sec": 0, 00:03:21.155 "rw_mbytes_per_sec": 0, 00:03:21.155 "r_mbytes_per_sec": 0, 00:03:21.155 "w_mbytes_per_sec": 0 00:03:21.155 }, 00:03:21.155 "claimed": false, 00:03:21.155 "zoned": false, 00:03:21.155 "supported_io_types": { 00:03:21.155 "read": true, 00:03:21.155 "write": true, 00:03:21.155 "unmap": true, 00:03:21.155 "flush": true, 00:03:21.155 "reset": true, 00:03:21.155 "nvme_admin": false, 00:03:21.155 "nvme_io": false, 00:03:21.155 "nvme_io_md": false, 00:03:21.155 "write_zeroes": true, 00:03:21.155 "zcopy": true, 00:03:21.155 "get_zone_info": false, 00:03:21.155 "zone_management": false, 00:03:21.155 "zone_append": false, 00:03:21.155 "compare": false, 00:03:21.155 "compare_and_write": false, 00:03:21.155 "abort": true, 00:03:21.155 "seek_hole": false, 00:03:21.155 "seek_data": false, 00:03:21.155 "copy": true, 00:03:21.155 "nvme_iov_md": false 00:03:21.155 }, 00:03:21.155 "memory_domains": [ 00:03:21.155 { 00:03:21.155 "dma_device_id": "system", 00:03:21.155 "dma_device_type": 1 00:03:21.155 }, 00:03:21.155 { 00:03:21.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:03:21.155 "dma_device_type": 2 00:03:21.155 } 00:03:21.155 ], 00:03:21.155 "driver_specific": {} 00:03:21.155 } 00:03:21.155 ]' 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.155 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.155 [2024-10-17 17:24:29.071943] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:03:21.155 [2024-10-17 17:24:29.071988] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:03:21.155 [2024-10-17 17:24:29.072003] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x16794b0 00:03:21.155 [2024-10-17 17:24:29.072011] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:03:21.155 [2024-10-17 17:24:29.073459] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:03:21.155 [2024-10-17 17:24:29.073503] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:03:21.417 Passthru0 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:03:21.417 { 00:03:21.417 "name": "Malloc2", 00:03:21.417 "aliases": [ 00:03:21.417 "309ce199-ea2c-4cce-985b-144b698da108" 00:03:21.417 ], 00:03:21.417 "product_name": "Malloc disk", 00:03:21.417 "block_size": 512, 00:03:21.417 "num_blocks": 16384, 00:03:21.417 "uuid": "309ce199-ea2c-4cce-985b-144b698da108", 00:03:21.417 "assigned_rate_limits": { 00:03:21.417 "rw_ios_per_sec": 0, 00:03:21.417 "rw_mbytes_per_sec": 0, 00:03:21.417 "r_mbytes_per_sec": 0, 00:03:21.417 "w_mbytes_per_sec": 0 00:03:21.417 }, 00:03:21.417 "claimed": true, 00:03:21.417 "claim_type": "exclusive_write", 00:03:21.417 "zoned": false, 00:03:21.417 "supported_io_types": { 00:03:21.417 "read": true, 00:03:21.417 "write": true, 00:03:21.417 "unmap": true, 00:03:21.417 "flush": true, 00:03:21.417 "reset": true, 00:03:21.417 "nvme_admin": false, 00:03:21.417 "nvme_io": false, 00:03:21.417 "nvme_io_md": false, 00:03:21.417 "write_zeroes": true, 00:03:21.417 "zcopy": true, 00:03:21.417 "get_zone_info": false, 00:03:21.417 "zone_management": false, 00:03:21.417 "zone_append": false, 00:03:21.417 "compare": false, 00:03:21.417 "compare_and_write": false, 00:03:21.417 "abort": true, 00:03:21.417 "seek_hole": false, 00:03:21.417 "seek_data": false, 00:03:21.417 "copy": true, 00:03:21.417 "nvme_iov_md": false 00:03:21.417 }, 00:03:21.417 "memory_domains": [ 00:03:21.417 { 00:03:21.417 "dma_device_id": "system", 00:03:21.417 "dma_device_type": 1 00:03:21.417 }, 00:03:21.417 { 00:03:21.417 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:03:21.417 "dma_device_type": 2 00:03:21.417 } 00:03:21.417 ], 00:03:21.417 "driver_specific": {} 00:03:21.417 }, 00:03:21.417 { 00:03:21.417 "name": "Passthru0", 00:03:21.417 "aliases": [ 00:03:21.417 "9e4d5cf7-7b17-567b-8af0-2db7d4ebbdc5" 00:03:21.417 ], 00:03:21.417 "product_name": "passthru", 00:03:21.417 "block_size": 512, 00:03:21.417 "num_blocks": 16384, 00:03:21.417 "uuid": "9e4d5cf7-7b17-567b-8af0-2db7d4ebbdc5", 00:03:21.417 "assigned_rate_limits": { 00:03:21.417 "rw_ios_per_sec": 0, 00:03:21.417 "rw_mbytes_per_sec": 0, 00:03:21.417 "r_mbytes_per_sec": 0, 00:03:21.417 "w_mbytes_per_sec": 0 00:03:21.417 }, 00:03:21.417 "claimed": false, 00:03:21.417 "zoned": false, 00:03:21.417 "supported_io_types": { 00:03:21.417 "read": true, 00:03:21.417 "write": true, 00:03:21.417 "unmap": true, 00:03:21.417 "flush": true, 00:03:21.417 "reset": true, 00:03:21.417 "nvme_admin": false, 00:03:21.417 "nvme_io": false, 00:03:21.417 "nvme_io_md": false, 00:03:21.417 "write_zeroes": true, 00:03:21.417 "zcopy": true, 00:03:21.417 "get_zone_info": false, 00:03:21.417 "zone_management": false, 00:03:21.417 "zone_append": false, 00:03:21.417 "compare": false, 00:03:21.417 "compare_and_write": false, 00:03:21.417 "abort": true, 00:03:21.417 "seek_hole": false, 00:03:21.417 "seek_data": false, 00:03:21.417 "copy": true, 00:03:21.417 "nvme_iov_md": false 00:03:21.417 }, 00:03:21.417 "memory_domains": [ 00:03:21.417 { 00:03:21.417 "dma_device_id": "system", 00:03:21.417 "dma_device_type": 1 00:03:21.417 }, 00:03:21.417 { 00:03:21.417 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:03:21.417 "dma_device_type": 2 00:03:21.417 } 00:03:21.417 ], 00:03:21.417 "driver_specific": { 00:03:21.417 "passthru": { 00:03:21.417 "name": "Passthru0", 00:03:21.417 "base_bdev_name": "Malloc2" 00:03:21.417 } 00:03:21.417 } 00:03:21.417 } 00:03:21.417 ]' 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:03:21.417 00:03:21.417 real 0m0.299s 00:03:21.417 user 0m0.195s 00:03:21.417 sys 0m0.040s 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:21.417 17:24:29 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:03:21.417 ************************************ 00:03:21.417 END TEST rpc_daemon_integrity 00:03:21.417 ************************************ 00:03:21.417 17:24:29 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:03:21.417 17:24:29 rpc -- rpc/rpc.sh@84 -- # killprocess 3998328 00:03:21.417 17:24:29 rpc -- common/autotest_common.sh@950 -- # '[' -z 3998328 ']' 00:03:21.417 17:24:29 rpc -- common/autotest_common.sh@954 -- # kill -0 3998328 00:03:21.417 17:24:29 rpc -- common/autotest_common.sh@955 -- # uname 00:03:21.417 17:24:29 rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:21.417 17:24:29 rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 3998328 00:03:21.678 17:24:29 rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:21.678 17:24:29 rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:21.678 17:24:29 rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 3998328' 00:03:21.678 killing process with pid 3998328 00:03:21.678 17:24:29 rpc -- common/autotest_common.sh@969 -- # kill 3998328 00:03:21.678 17:24:29 rpc -- common/autotest_common.sh@974 -- # wait 3998328 00:03:21.678 00:03:21.678 real 0m2.713s 00:03:21.678 user 0m3.444s 00:03:21.678 sys 0m0.856s 00:03:21.678 17:24:29 rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:21.678 17:24:29 rpc -- common/autotest_common.sh@10 -- # set +x 00:03:21.678 ************************************ 00:03:21.678 END TEST rpc 00:03:21.678 ************************************ 00:03:21.939 17:24:29 -- spdk/autotest.sh@157 -- # run_test skip_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:03:21.939 17:24:29 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:21.939 17:24:29 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:21.939 17:24:29 -- common/autotest_common.sh@10 -- # set +x 00:03:21.939 ************************************ 00:03:21.939 START TEST skip_rpc 00:03:21.939 ************************************ 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:03:21.939 * Looking for test storage... 00:03:21.939 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1691 -- # lcov --version 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@345 -- # : 1 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:21.939 17:24:29 skip_rpc -- scripts/common.sh@368 -- # return 0 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:21.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:21.939 --rc genhtml_branch_coverage=1 00:03:21.939 --rc genhtml_function_coverage=1 00:03:21.939 --rc genhtml_legend=1 00:03:21.939 --rc geninfo_all_blocks=1 00:03:21.939 --rc geninfo_unexecuted_blocks=1 00:03:21.939 00:03:21.939 ' 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:21.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:21.939 --rc genhtml_branch_coverage=1 00:03:21.939 --rc genhtml_function_coverage=1 00:03:21.939 --rc genhtml_legend=1 00:03:21.939 --rc geninfo_all_blocks=1 00:03:21.939 --rc geninfo_unexecuted_blocks=1 00:03:21.939 00:03:21.939 ' 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:21.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:21.939 --rc genhtml_branch_coverage=1 00:03:21.939 --rc genhtml_function_coverage=1 00:03:21.939 --rc genhtml_legend=1 00:03:21.939 --rc geninfo_all_blocks=1 00:03:21.939 --rc geninfo_unexecuted_blocks=1 00:03:21.939 00:03:21.939 ' 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:21.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:21.939 --rc genhtml_branch_coverage=1 00:03:21.939 --rc genhtml_function_coverage=1 00:03:21.939 --rc genhtml_legend=1 00:03:21.939 --rc geninfo_all_blocks=1 00:03:21.939 --rc geninfo_unexecuted_blocks=1 00:03:21.939 00:03:21.939 ' 00:03:21.939 17:24:29 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:03:21.939 17:24:29 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:03:21.939 17:24:29 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:03:21.939 17:24:29 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:22.200 17:24:29 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:22.200 17:24:29 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:22.200 ************************************ 00:03:22.200 START TEST skip_rpc 00:03:22.200 ************************************ 00:03:22.200 17:24:29 skip_rpc.skip_rpc -- common/autotest_common.sh@1125 -- # test_skip_rpc 00:03:22.200 17:24:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=3999177 00:03:22.200 17:24:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:03:22.201 17:24:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:03:22.201 17:24:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:03:22.201 [2024-10-17 17:24:29.955980] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:22.201 [2024-10-17 17:24:29.956039] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid3999177 ] 00:03:22.201 [2024-10-17 17:24:30.043970] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:22.201 [2024-10-17 17:24:30.100553] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # local es=0 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd spdk_get_version 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # rpc_cmd spdk_get_version 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # es=1 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 3999177 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@950 -- # '[' -z 3999177 ']' 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # kill -0 3999177 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # uname 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 3999177 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 3999177' 00:03:27.493 killing process with pid 3999177 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@969 -- # kill 3999177 00:03:27.493 17:24:34 skip_rpc.skip_rpc -- common/autotest_common.sh@974 -- # wait 3999177 00:03:27.493 00:03:27.493 real 0m5.262s 00:03:27.493 user 0m5.002s 00:03:27.493 sys 0m0.306s 00:03:27.493 17:24:35 skip_rpc.skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:27.493 17:24:35 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:27.493 ************************************ 00:03:27.493 END TEST skip_rpc 00:03:27.493 ************************************ 00:03:27.493 17:24:35 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:03:27.493 17:24:35 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:27.493 17:24:35 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:27.493 17:24:35 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:27.493 ************************************ 00:03:27.493 START TEST skip_rpc_with_json 00:03:27.493 ************************************ 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_json 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=4000216 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 4000216 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@831 -- # '[' -z 4000216 ']' 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:03:27.493 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:27.493 17:24:35 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:03:27.493 [2024-10-17 17:24:35.293955] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:27.493 [2024-10-17 17:24:35.294009] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4000216 ] 00:03:27.493 [2024-10-17 17:24:35.373866] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:27.493 [2024-10-17 17:24:35.408067] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # return 0 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:03:28.445 [2024-10-17 17:24:36.080275] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:03:28.445 request: 00:03:28.445 { 00:03:28.445 "trtype": "tcp", 00:03:28.445 "method": "nvmf_get_transports", 00:03:28.445 "req_id": 1 00:03:28.445 } 00:03:28.445 Got JSON-RPC error response 00:03:28.445 response: 00:03:28.445 { 00:03:28.445 "code": -19, 00:03:28.445 "message": "No such device" 00:03:28.445 } 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:03:28.445 [2024-10-17 17:24:36.092400] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:28.445 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:03:28.445 { 00:03:28.445 "subsystems": [ 00:03:28.445 { 00:03:28.445 "subsystem": "fsdev", 00:03:28.445 "config": [ 00:03:28.445 { 00:03:28.445 "method": "fsdev_set_opts", 00:03:28.445 "params": { 00:03:28.445 "fsdev_io_pool_size": 65535, 00:03:28.445 "fsdev_io_cache_size": 256 00:03:28.445 } 00:03:28.445 } 00:03:28.445 ] 00:03:28.445 }, 00:03:28.445 { 00:03:28.445 "subsystem": "vfio_user_target", 00:03:28.445 "config": null 00:03:28.445 }, 00:03:28.445 { 00:03:28.445 "subsystem": "keyring", 00:03:28.445 "config": [] 00:03:28.445 }, 00:03:28.445 { 00:03:28.445 "subsystem": "iobuf", 00:03:28.445 "config": [ 00:03:28.445 { 00:03:28.445 "method": "iobuf_set_options", 00:03:28.445 "params": { 00:03:28.445 "small_pool_count": 8192, 00:03:28.445 "large_pool_count": 1024, 00:03:28.445 "small_bufsize": 8192, 00:03:28.445 "large_bufsize": 135168 00:03:28.445 } 00:03:28.445 } 00:03:28.445 ] 00:03:28.445 }, 00:03:28.445 { 00:03:28.445 "subsystem": "sock", 00:03:28.445 "config": [ 00:03:28.445 { 00:03:28.445 "method": "sock_set_default_impl", 00:03:28.445 "params": { 00:03:28.445 "impl_name": "posix" 00:03:28.445 } 00:03:28.445 }, 00:03:28.445 { 00:03:28.446 "method": "sock_impl_set_options", 00:03:28.446 "params": { 00:03:28.446 "impl_name": "ssl", 00:03:28.446 "recv_buf_size": 4096, 00:03:28.446 "send_buf_size": 4096, 00:03:28.446 "enable_recv_pipe": true, 00:03:28.446 "enable_quickack": false, 00:03:28.446 "enable_placement_id": 0, 00:03:28.446 "enable_zerocopy_send_server": true, 00:03:28.446 "enable_zerocopy_send_client": false, 00:03:28.446 "zerocopy_threshold": 0, 00:03:28.446 "tls_version": 0, 00:03:28.446 "enable_ktls": false 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "sock_impl_set_options", 00:03:28.446 "params": { 00:03:28.446 "impl_name": "posix", 00:03:28.446 "recv_buf_size": 2097152, 00:03:28.446 "send_buf_size": 2097152, 00:03:28.446 "enable_recv_pipe": true, 00:03:28.446 "enable_quickack": false, 00:03:28.446 "enable_placement_id": 0, 00:03:28.446 "enable_zerocopy_send_server": true, 00:03:28.446 "enable_zerocopy_send_client": false, 00:03:28.446 "zerocopy_threshold": 0, 00:03:28.446 "tls_version": 0, 00:03:28.446 "enable_ktls": false 00:03:28.446 } 00:03:28.446 } 00:03:28.446 ] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "vmd", 00:03:28.446 "config": [] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "accel", 00:03:28.446 "config": [ 00:03:28.446 { 00:03:28.446 "method": "accel_set_options", 00:03:28.446 "params": { 00:03:28.446 "small_cache_size": 128, 00:03:28.446 "large_cache_size": 16, 00:03:28.446 "task_count": 2048, 00:03:28.446 "sequence_count": 2048, 00:03:28.446 "buf_count": 2048 00:03:28.446 } 00:03:28.446 } 00:03:28.446 ] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "bdev", 00:03:28.446 "config": [ 00:03:28.446 { 00:03:28.446 "method": "bdev_set_options", 00:03:28.446 "params": { 00:03:28.446 "bdev_io_pool_size": 65535, 00:03:28.446 "bdev_io_cache_size": 256, 00:03:28.446 "bdev_auto_examine": true, 00:03:28.446 "iobuf_small_cache_size": 128, 00:03:28.446 "iobuf_large_cache_size": 16 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "bdev_raid_set_options", 00:03:28.446 "params": { 00:03:28.446 "process_window_size_kb": 1024, 00:03:28.446 "process_max_bandwidth_mb_sec": 0 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "bdev_iscsi_set_options", 00:03:28.446 "params": { 00:03:28.446 "timeout_sec": 30 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "bdev_nvme_set_options", 00:03:28.446 "params": { 00:03:28.446 "action_on_timeout": "none", 00:03:28.446 "timeout_us": 0, 00:03:28.446 "timeout_admin_us": 0, 00:03:28.446 "keep_alive_timeout_ms": 10000, 00:03:28.446 "arbitration_burst": 0, 00:03:28.446 "low_priority_weight": 0, 00:03:28.446 "medium_priority_weight": 0, 00:03:28.446 "high_priority_weight": 0, 00:03:28.446 "nvme_adminq_poll_period_us": 10000, 00:03:28.446 "nvme_ioq_poll_period_us": 0, 00:03:28.446 "io_queue_requests": 0, 00:03:28.446 "delay_cmd_submit": true, 00:03:28.446 "transport_retry_count": 4, 00:03:28.446 "bdev_retry_count": 3, 00:03:28.446 "transport_ack_timeout": 0, 00:03:28.446 "ctrlr_loss_timeout_sec": 0, 00:03:28.446 "reconnect_delay_sec": 0, 00:03:28.446 "fast_io_fail_timeout_sec": 0, 00:03:28.446 "disable_auto_failback": false, 00:03:28.446 "generate_uuids": false, 00:03:28.446 "transport_tos": 0, 00:03:28.446 "nvme_error_stat": false, 00:03:28.446 "rdma_srq_size": 0, 00:03:28.446 "io_path_stat": false, 00:03:28.446 "allow_accel_sequence": false, 00:03:28.446 "rdma_max_cq_size": 0, 00:03:28.446 "rdma_cm_event_timeout_ms": 0, 00:03:28.446 "dhchap_digests": [ 00:03:28.446 "sha256", 00:03:28.446 "sha384", 00:03:28.446 "sha512" 00:03:28.446 ], 00:03:28.446 "dhchap_dhgroups": [ 00:03:28.446 "null", 00:03:28.446 "ffdhe2048", 00:03:28.446 "ffdhe3072", 00:03:28.446 "ffdhe4096", 00:03:28.446 "ffdhe6144", 00:03:28.446 "ffdhe8192" 00:03:28.446 ], 00:03:28.446 "rdma_umr_per_io": false 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "bdev_nvme_set_hotplug", 00:03:28.446 "params": { 00:03:28.446 "period_us": 100000, 00:03:28.446 "enable": false 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "bdev_wait_for_examine" 00:03:28.446 } 00:03:28.446 ] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "scsi", 00:03:28.446 "config": null 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "scheduler", 00:03:28.446 "config": [ 00:03:28.446 { 00:03:28.446 "method": "framework_set_scheduler", 00:03:28.446 "params": { 00:03:28.446 "name": "static" 00:03:28.446 } 00:03:28.446 } 00:03:28.446 ] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "vhost_scsi", 00:03:28.446 "config": [] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "vhost_blk", 00:03:28.446 "config": [] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "ublk", 00:03:28.446 "config": [] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "nbd", 00:03:28.446 "config": [] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "nvmf", 00:03:28.446 "config": [ 00:03:28.446 { 00:03:28.446 "method": "nvmf_set_config", 00:03:28.446 "params": { 00:03:28.446 "discovery_filter": "match_any", 00:03:28.446 "admin_cmd_passthru": { 00:03:28.446 "identify_ctrlr": false 00:03:28.446 }, 00:03:28.446 "dhchap_digests": [ 00:03:28.446 "sha256", 00:03:28.446 "sha384", 00:03:28.446 "sha512" 00:03:28.446 ], 00:03:28.446 "dhchap_dhgroups": [ 00:03:28.446 "null", 00:03:28.446 "ffdhe2048", 00:03:28.446 "ffdhe3072", 00:03:28.446 "ffdhe4096", 00:03:28.446 "ffdhe6144", 00:03:28.446 "ffdhe8192" 00:03:28.446 ] 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "nvmf_set_max_subsystems", 00:03:28.446 "params": { 00:03:28.446 "max_subsystems": 1024 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "nvmf_set_crdt", 00:03:28.446 "params": { 00:03:28.446 "crdt1": 0, 00:03:28.446 "crdt2": 0, 00:03:28.446 "crdt3": 0 00:03:28.446 } 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "method": "nvmf_create_transport", 00:03:28.446 "params": { 00:03:28.446 "trtype": "TCP", 00:03:28.446 "max_queue_depth": 128, 00:03:28.446 "max_io_qpairs_per_ctrlr": 127, 00:03:28.446 "in_capsule_data_size": 4096, 00:03:28.446 "max_io_size": 131072, 00:03:28.446 "io_unit_size": 131072, 00:03:28.446 "max_aq_depth": 128, 00:03:28.446 "num_shared_buffers": 511, 00:03:28.446 "buf_cache_size": 4294967295, 00:03:28.446 "dif_insert_or_strip": false, 00:03:28.446 "zcopy": false, 00:03:28.446 "c2h_success": true, 00:03:28.446 "sock_priority": 0, 00:03:28.446 "abort_timeout_sec": 1, 00:03:28.446 "ack_timeout": 0, 00:03:28.446 "data_wr_pool_size": 0 00:03:28.446 } 00:03:28.446 } 00:03:28.446 ] 00:03:28.446 }, 00:03:28.446 { 00:03:28.446 "subsystem": "iscsi", 00:03:28.446 "config": [ 00:03:28.446 { 00:03:28.446 "method": "iscsi_set_options", 00:03:28.446 "params": { 00:03:28.446 "node_base": "iqn.2016-06.io.spdk", 00:03:28.446 "max_sessions": 128, 00:03:28.446 "max_connections_per_session": 2, 00:03:28.446 "max_queue_depth": 64, 00:03:28.446 "default_time2wait": 2, 00:03:28.446 "default_time2retain": 20, 00:03:28.446 "first_burst_length": 8192, 00:03:28.446 "immediate_data": true, 00:03:28.446 "allow_duplicated_isid": false, 00:03:28.446 "error_recovery_level": 0, 00:03:28.446 "nop_timeout": 60, 00:03:28.446 "nop_in_interval": 30, 00:03:28.446 "disable_chap": false, 00:03:28.446 "require_chap": false, 00:03:28.446 "mutual_chap": false, 00:03:28.446 "chap_group": 0, 00:03:28.446 "max_large_datain_per_connection": 64, 00:03:28.447 "max_r2t_per_connection": 4, 00:03:28.447 "pdu_pool_size": 36864, 00:03:28.447 "immediate_data_pool_size": 16384, 00:03:28.447 "data_out_pool_size": 2048 00:03:28.447 } 00:03:28.447 } 00:03:28.447 ] 00:03:28.447 } 00:03:28.447 ] 00:03:28.447 } 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 4000216 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 4000216 ']' 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 4000216 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4000216 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4000216' 00:03:28.447 killing process with pid 4000216 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 4000216 00:03:28.447 17:24:36 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 4000216 00:03:28.707 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=4000556 00:03:28.707 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:03:28.707 17:24:36 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 4000556 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 4000556 ']' 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 4000556 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4000556 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4000556' 00:03:33.984 killing process with pid 4000556 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 4000556 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 4000556 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:03:33.984 00:03:33.984 real 0m6.543s 00:03:33.984 user 0m6.431s 00:03:33.984 sys 0m0.571s 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:33.984 17:24:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:03:33.984 ************************************ 00:03:33.984 END TEST skip_rpc_with_json 00:03:33.984 ************************************ 00:03:33.985 17:24:41 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:03:33.985 17:24:41 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:33.985 17:24:41 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:33.985 17:24:41 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:33.985 ************************************ 00:03:33.985 START TEST skip_rpc_with_delay 00:03:33.985 ************************************ 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_delay 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # local es=0 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:03:33.985 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:03:34.246 [2024-10-17 17:24:41.916303] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:03:34.246 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # es=1 00:03:34.246 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:03:34.246 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:03:34.246 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:03:34.246 00:03:34.246 real 0m0.075s 00:03:34.246 user 0m0.046s 00:03:34.246 sys 0m0.029s 00:03:34.246 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:34.246 17:24:41 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:03:34.246 ************************************ 00:03:34.246 END TEST skip_rpc_with_delay 00:03:34.246 ************************************ 00:03:34.246 17:24:41 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:03:34.246 17:24:41 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:03:34.246 17:24:41 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:03:34.246 17:24:41 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:34.246 17:24:41 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:34.246 17:24:41 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:34.246 ************************************ 00:03:34.246 START TEST exit_on_failed_rpc_init 00:03:34.246 ************************************ 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1125 -- # test_exit_on_failed_rpc_init 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=4001623 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 4001623 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@831 -- # '[' -z 4001623 ']' 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:03:34.246 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:34.246 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:03:34.246 [2024-10-17 17:24:42.074536] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:34.246 [2024-10-17 17:24:42.074597] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4001623 ] 00:03:34.246 [2024-10-17 17:24:42.155028] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:34.507 [2024-10-17 17:24:42.190864] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # return 0 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # local es=0 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:03:35.078 17:24:42 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:03:35.078 [2024-10-17 17:24:42.942407] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:35.078 [2024-10-17 17:24:42.942459] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4001806 ] 00:03:35.338 [2024-10-17 17:24:43.021564] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:35.338 [2024-10-17 17:24:43.057512] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:03:35.338 [2024-10-17 17:24:43.057559] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:03:35.338 [2024-10-17 17:24:43.057569] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:03:35.338 [2024-10-17 17:24:43.057576] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # es=234 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@662 -- # es=106 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # case "$es" in 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@670 -- # es=1 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 4001623 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@950 -- # '[' -z 4001623 ']' 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # kill -0 4001623 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # uname 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4001623 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4001623' 00:03:35.338 killing process with pid 4001623 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@969 -- # kill 4001623 00:03:35.338 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@974 -- # wait 4001623 00:03:35.598 00:03:35.598 real 0m1.331s 00:03:35.598 user 0m1.556s 00:03:35.598 sys 0m0.400s 00:03:35.598 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:35.598 17:24:43 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:03:35.598 ************************************ 00:03:35.598 END TEST exit_on_failed_rpc_init 00:03:35.598 ************************************ 00:03:35.598 17:24:43 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:03:35.598 00:03:35.598 real 0m13.731s 00:03:35.598 user 0m13.242s 00:03:35.598 sys 0m1.647s 00:03:35.598 17:24:43 skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:35.598 17:24:43 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:35.598 ************************************ 00:03:35.598 END TEST skip_rpc 00:03:35.598 ************************************ 00:03:35.598 17:24:43 -- spdk/autotest.sh@158 -- # run_test rpc_client /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:03:35.598 17:24:43 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:35.598 17:24:43 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:35.598 17:24:43 -- common/autotest_common.sh@10 -- # set +x 00:03:35.598 ************************************ 00:03:35.598 START TEST rpc_client 00:03:35.598 ************************************ 00:03:35.598 17:24:43 rpc_client -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:03:35.859 * Looking for test storage... 00:03:35.859 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1691 -- # lcov --version 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@345 -- # : 1 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@353 -- # local d=1 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@355 -- # echo 1 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@353 -- # local d=2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@355 -- # echo 2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:35.859 17:24:43 rpc_client -- scripts/common.sh@368 -- # return 0 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:35.859 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:35.859 --rc genhtml_branch_coverage=1 00:03:35.859 --rc genhtml_function_coverage=1 00:03:35.859 --rc genhtml_legend=1 00:03:35.859 --rc geninfo_all_blocks=1 00:03:35.859 --rc geninfo_unexecuted_blocks=1 00:03:35.859 00:03:35.859 ' 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:35.859 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:35.859 --rc genhtml_branch_coverage=1 00:03:35.859 --rc genhtml_function_coverage=1 00:03:35.859 --rc genhtml_legend=1 00:03:35.859 --rc geninfo_all_blocks=1 00:03:35.859 --rc geninfo_unexecuted_blocks=1 00:03:35.859 00:03:35.859 ' 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:35.859 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:35.859 --rc genhtml_branch_coverage=1 00:03:35.859 --rc genhtml_function_coverage=1 00:03:35.859 --rc genhtml_legend=1 00:03:35.859 --rc geninfo_all_blocks=1 00:03:35.859 --rc geninfo_unexecuted_blocks=1 00:03:35.859 00:03:35.859 ' 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:35.859 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:35.859 --rc genhtml_branch_coverage=1 00:03:35.859 --rc genhtml_function_coverage=1 00:03:35.859 --rc genhtml_legend=1 00:03:35.859 --rc geninfo_all_blocks=1 00:03:35.859 --rc geninfo_unexecuted_blocks=1 00:03:35.859 00:03:35.859 ' 00:03:35.859 17:24:43 rpc_client -- rpc_client/rpc_client.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client_test 00:03:35.859 OK 00:03:35.859 17:24:43 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:03:35.859 00:03:35.859 real 0m0.223s 00:03:35.859 user 0m0.134s 00:03:35.859 sys 0m0.101s 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:35.859 17:24:43 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:03:35.859 ************************************ 00:03:35.859 END TEST rpc_client 00:03:35.859 ************************************ 00:03:35.859 17:24:43 -- spdk/autotest.sh@159 -- # run_test json_config /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:03:35.859 17:24:43 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:35.859 17:24:43 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:35.859 17:24:43 -- common/autotest_common.sh@10 -- # set +x 00:03:35.859 ************************************ 00:03:35.859 START TEST json_config 00:03:35.859 ************************************ 00:03:35.860 17:24:43 json_config -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:03:36.119 17:24:43 json_config -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:36.119 17:24:43 json_config -- common/autotest_common.sh@1691 -- # lcov --version 00:03:36.119 17:24:43 json_config -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:36.119 17:24:43 json_config -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:36.119 17:24:43 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:36.120 17:24:43 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:36.120 17:24:43 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:36.120 17:24:43 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:03:36.120 17:24:43 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:03:36.120 17:24:43 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:03:36.120 17:24:43 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:03:36.120 17:24:43 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:03:36.120 17:24:43 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:03:36.120 17:24:43 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:03:36.120 17:24:43 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:36.120 17:24:43 json_config -- scripts/common.sh@344 -- # case "$op" in 00:03:36.120 17:24:43 json_config -- scripts/common.sh@345 -- # : 1 00:03:36.120 17:24:43 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:36.120 17:24:43 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:36.120 17:24:43 json_config -- scripts/common.sh@365 -- # decimal 1 00:03:36.120 17:24:43 json_config -- scripts/common.sh@353 -- # local d=1 00:03:36.120 17:24:43 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:36.120 17:24:43 json_config -- scripts/common.sh@355 -- # echo 1 00:03:36.120 17:24:43 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:03:36.120 17:24:43 json_config -- scripts/common.sh@366 -- # decimal 2 00:03:36.120 17:24:43 json_config -- scripts/common.sh@353 -- # local d=2 00:03:36.120 17:24:43 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:36.120 17:24:43 json_config -- scripts/common.sh@355 -- # echo 2 00:03:36.120 17:24:43 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:03:36.120 17:24:43 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:36.120 17:24:43 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:36.120 17:24:43 json_config -- scripts/common.sh@368 -- # return 0 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:36.120 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:36.120 --rc genhtml_branch_coverage=1 00:03:36.120 --rc genhtml_function_coverage=1 00:03:36.120 --rc genhtml_legend=1 00:03:36.120 --rc geninfo_all_blocks=1 00:03:36.120 --rc geninfo_unexecuted_blocks=1 00:03:36.120 00:03:36.120 ' 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:36.120 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:36.120 --rc genhtml_branch_coverage=1 00:03:36.120 --rc genhtml_function_coverage=1 00:03:36.120 --rc genhtml_legend=1 00:03:36.120 --rc geninfo_all_blocks=1 00:03:36.120 --rc geninfo_unexecuted_blocks=1 00:03:36.120 00:03:36.120 ' 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:36.120 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:36.120 --rc genhtml_branch_coverage=1 00:03:36.120 --rc genhtml_function_coverage=1 00:03:36.120 --rc genhtml_legend=1 00:03:36.120 --rc geninfo_all_blocks=1 00:03:36.120 --rc geninfo_unexecuted_blocks=1 00:03:36.120 00:03:36.120 ' 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:36.120 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:36.120 --rc genhtml_branch_coverage=1 00:03:36.120 --rc genhtml_function_coverage=1 00:03:36.120 --rc genhtml_legend=1 00:03:36.120 --rc geninfo_all_blocks=1 00:03:36.120 --rc geninfo_unexecuted_blocks=1 00:03:36.120 00:03:36.120 ' 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@7 -- # uname -s 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:03:36.120 17:24:43 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:03:36.120 17:24:43 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:03:36.120 17:24:43 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:03:36.120 17:24:43 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:03:36.120 17:24:43 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:36.120 17:24:43 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:36.120 17:24:43 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:36.120 17:24:43 json_config -- paths/export.sh@5 -- # export PATH 00:03:36.120 17:24:43 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@51 -- # : 0 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:03:36.120 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:03:36.120 17:24:43 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@31 -- # app_pid=(['target']='' ['initiator']='') 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@31 -- # declare -A app_pid 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@32 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock' ['initiator']='/var/tmp/spdk_initiator.sock') 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@32 -- # declare -A app_socket 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@33 -- # app_params=(['target']='-m 0x1 -s 1024' ['initiator']='-m 0x2 -g -u -s 1024') 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@33 -- # declare -A app_params 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@34 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json' ['initiator']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json') 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@34 -- # declare -A configs_path 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@40 -- # last_event_id=0 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@362 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@363 -- # echo 'INFO: JSON configuration test init' 00:03:36.120 INFO: JSON configuration test init 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@364 -- # json_config_test_init 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@269 -- # timing_enter json_config_test_init 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@270 -- # timing_enter json_config_setup_target 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:36.120 17:24:43 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:36.120 17:24:43 json_config -- json_config/json_config.sh@272 -- # json_config_test_start_app target --wait-for-rpc 00:03:36.120 17:24:43 json_config -- json_config/common.sh@9 -- # local app=target 00:03:36.120 17:24:43 json_config -- json_config/common.sh@10 -- # shift 00:03:36.120 17:24:43 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:03:36.120 17:24:43 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:03:36.120 17:24:43 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:03:36.120 17:24:43 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:03:36.120 17:24:43 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:03:36.120 17:24:43 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=4002094 00:03:36.120 17:24:43 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:03:36.120 Waiting for target to run... 00:03:36.120 17:24:43 json_config -- json_config/common.sh@25 -- # waitforlisten 4002094 /var/tmp/spdk_tgt.sock 00:03:36.121 17:24:43 json_config -- common/autotest_common.sh@831 -- # '[' -z 4002094 ']' 00:03:36.121 17:24:43 json_config -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:03:36.121 17:24:43 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --wait-for-rpc 00:03:36.121 17:24:43 json_config -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:36.121 17:24:43 json_config -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:03:36.121 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:03:36.121 17:24:43 json_config -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:36.121 17:24:43 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:36.380 [2024-10-17 17:24:44.054420] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:36.380 [2024-10-17 17:24:44.054475] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4002094 ] 00:03:36.642 [2024-10-17 17:24:44.356539] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:36.642 [2024-10-17 17:24:44.384140] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:37.212 17:24:44 json_config -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:37.212 17:24:44 json_config -- common/autotest_common.sh@864 -- # return 0 00:03:37.212 17:24:44 json_config -- json_config/common.sh@26 -- # echo '' 00:03:37.212 00:03:37.212 17:24:44 json_config -- json_config/json_config.sh@276 -- # create_accel_config 00:03:37.212 17:24:44 json_config -- json_config/json_config.sh@100 -- # timing_enter create_accel_config 00:03:37.212 17:24:44 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:37.212 17:24:44 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:37.212 17:24:44 json_config -- json_config/json_config.sh@102 -- # [[ 0 -eq 1 ]] 00:03:37.212 17:24:44 json_config -- json_config/json_config.sh@108 -- # timing_exit create_accel_config 00:03:37.212 17:24:44 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:37.212 17:24:44 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:37.212 17:24:44 json_config -- json_config/json_config.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh --json-with-subsystems 00:03:37.212 17:24:44 json_config -- json_config/json_config.sh@281 -- # tgt_rpc load_config 00:03:37.212 17:24:44 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock load_config 00:03:37.781 17:24:45 json_config -- json_config/json_config.sh@283 -- # tgt_check_notification_types 00:03:37.781 17:24:45 json_config -- json_config/json_config.sh@43 -- # timing_enter tgt_check_notification_types 00:03:37.781 17:24:45 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:37.781 17:24:45 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:37.781 17:24:45 json_config -- json_config/json_config.sh@45 -- # local ret=0 00:03:37.781 17:24:45 json_config -- json_config/json_config.sh@46 -- # enabled_types=('bdev_register' 'bdev_unregister') 00:03:37.781 17:24:45 json_config -- json_config/json_config.sh@46 -- # local enabled_types 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@47 -- # [[ y == y ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@48 -- # enabled_types+=("fsdev_register" "fsdev_unregister") 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@51 -- # tgt_rpc notify_get_types 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@51 -- # jq -r '.[]' 00:03:37.782 17:24:45 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_types 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@51 -- # get_types=('fsdev_register' 'fsdev_unregister' 'bdev_register' 'bdev_unregister') 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@51 -- # local get_types 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@53 -- # local type_diff 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@54 -- # echo bdev_register bdev_unregister fsdev_register fsdev_unregister fsdev_register fsdev_unregister bdev_register bdev_unregister 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@54 -- # tr ' ' '\n' 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@54 -- # sort 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@54 -- # uniq -u 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@54 -- # type_diff= 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@56 -- # [[ -n '' ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@61 -- # timing_exit tgt_check_notification_types 00:03:37.782 17:24:45 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:37.782 17:24:45 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@62 -- # return 0 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@285 -- # [[ 0 -eq 1 ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@289 -- # [[ 0 -eq 1 ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@293 -- # [[ 0 -eq 1 ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@297 -- # [[ 1 -eq 1 ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@298 -- # create_nvmf_subsystem_config 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@237 -- # timing_enter create_nvmf_subsystem_config 00:03:37.782 17:24:45 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:37.782 17:24:45 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@239 -- # NVMF_FIRST_TARGET_IP=127.0.0.1 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@240 -- # [[ tcp == \r\d\m\a ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@244 -- # [[ -z 127.0.0.1 ]] 00:03:37.782 17:24:45 json_config -- json_config/json_config.sh@249 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocForNvmf0 00:03:37.782 17:24:45 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocForNvmf0 00:03:38.043 MallocForNvmf0 00:03:38.043 17:24:45 json_config -- json_config/json_config.sh@250 -- # tgt_rpc bdev_malloc_create 4 1024 --name MallocForNvmf1 00:03:38.043 17:24:45 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 4 1024 --name MallocForNvmf1 00:03:38.304 MallocForNvmf1 00:03:38.304 17:24:46 json_config -- json_config/json_config.sh@252 -- # tgt_rpc nvmf_create_transport -t tcp -u 8192 -c 0 00:03:38.304 17:24:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_transport -t tcp -u 8192 -c 0 00:03:38.304 [2024-10-17 17:24:46.170004] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:03:38.304 17:24:46 json_config -- json_config/json_config.sh@253 -- # tgt_rpc nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:03:38.304 17:24:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:03:38.564 17:24:46 json_config -- json_config/json_config.sh@254 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:03:38.564 17:24:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:03:38.825 17:24:46 json_config -- json_config/json_config.sh@255 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:03:38.825 17:24:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:03:38.825 17:24:46 json_config -- json_config/json_config.sh@256 -- # tgt_rpc nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:03:38.825 17:24:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:03:39.086 [2024-10-17 17:24:46.799956] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:03:39.086 17:24:46 json_config -- json_config/json_config.sh@258 -- # timing_exit create_nvmf_subsystem_config 00:03:39.086 17:24:46 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:39.086 17:24:46 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:39.086 17:24:46 json_config -- json_config/json_config.sh@300 -- # timing_exit json_config_setup_target 00:03:39.086 17:24:46 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:39.086 17:24:46 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:39.086 17:24:46 json_config -- json_config/json_config.sh@302 -- # [[ 0 -eq 1 ]] 00:03:39.086 17:24:46 json_config -- json_config/json_config.sh@307 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:03:39.086 17:24:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:03:39.347 MallocBdevForConfigChangeCheck 00:03:39.347 17:24:47 json_config -- json_config/json_config.sh@309 -- # timing_exit json_config_test_init 00:03:39.347 17:24:47 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:39.347 17:24:47 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:39.347 17:24:47 json_config -- json_config/json_config.sh@366 -- # tgt_rpc save_config 00:03:39.347 17:24:47 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:03:39.606 17:24:47 json_config -- json_config/json_config.sh@368 -- # echo 'INFO: shutting down applications...' 00:03:39.606 INFO: shutting down applications... 00:03:39.606 17:24:47 json_config -- json_config/json_config.sh@369 -- # [[ 0 -eq 1 ]] 00:03:39.606 17:24:47 json_config -- json_config/json_config.sh@375 -- # json_config_clear target 00:03:39.606 17:24:47 json_config -- json_config/json_config.sh@339 -- # [[ -n 22 ]] 00:03:39.606 17:24:47 json_config -- json_config/json_config.sh@340 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py -s /var/tmp/spdk_tgt.sock clear_config 00:03:40.177 Calling clear_iscsi_subsystem 00:03:40.177 Calling clear_nvmf_subsystem 00:03:40.177 Calling clear_nbd_subsystem 00:03:40.177 Calling clear_ublk_subsystem 00:03:40.177 Calling clear_vhost_blk_subsystem 00:03:40.177 Calling clear_vhost_scsi_subsystem 00:03:40.177 Calling clear_bdev_subsystem 00:03:40.177 17:24:47 json_config -- json_config/json_config.sh@344 -- # local config_filter=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py 00:03:40.177 17:24:47 json_config -- json_config/json_config.sh@350 -- # count=100 00:03:40.177 17:24:47 json_config -- json_config/json_config.sh@351 -- # '[' 100 -gt 0 ']' 00:03:40.177 17:24:47 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:03:40.177 17:24:47 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method delete_global_parameters 00:03:40.177 17:24:47 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method check_empty 00:03:40.438 17:24:48 json_config -- json_config/json_config.sh@352 -- # break 00:03:40.438 17:24:48 json_config -- json_config/json_config.sh@357 -- # '[' 100 -eq 0 ']' 00:03:40.438 17:24:48 json_config -- json_config/json_config.sh@376 -- # json_config_test_shutdown_app target 00:03:40.438 17:24:48 json_config -- json_config/common.sh@31 -- # local app=target 00:03:40.438 17:24:48 json_config -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:03:40.438 17:24:48 json_config -- json_config/common.sh@35 -- # [[ -n 4002094 ]] 00:03:40.438 17:24:48 json_config -- json_config/common.sh@38 -- # kill -SIGINT 4002094 00:03:40.438 17:24:48 json_config -- json_config/common.sh@40 -- # (( i = 0 )) 00:03:40.438 17:24:48 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:03:40.438 17:24:48 json_config -- json_config/common.sh@41 -- # kill -0 4002094 00:03:40.438 17:24:48 json_config -- json_config/common.sh@45 -- # sleep 0.5 00:03:41.010 17:24:48 json_config -- json_config/common.sh@40 -- # (( i++ )) 00:03:41.010 17:24:48 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:03:41.010 17:24:48 json_config -- json_config/common.sh@41 -- # kill -0 4002094 00:03:41.010 17:24:48 json_config -- json_config/common.sh@42 -- # app_pid["$app"]= 00:03:41.010 17:24:48 json_config -- json_config/common.sh@43 -- # break 00:03:41.010 17:24:48 json_config -- json_config/common.sh@48 -- # [[ -n '' ]] 00:03:41.010 17:24:48 json_config -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:03:41.010 SPDK target shutdown done 00:03:41.010 17:24:48 json_config -- json_config/json_config.sh@378 -- # echo 'INFO: relaunching applications...' 00:03:41.010 INFO: relaunching applications... 00:03:41.010 17:24:48 json_config -- json_config/json_config.sh@379 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:03:41.010 17:24:48 json_config -- json_config/common.sh@9 -- # local app=target 00:03:41.010 17:24:48 json_config -- json_config/common.sh@10 -- # shift 00:03:41.010 17:24:48 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:03:41.010 17:24:48 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:03:41.010 17:24:48 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:03:41.010 17:24:48 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:03:41.010 17:24:48 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:03:41.010 17:24:48 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=4003230 00:03:41.010 17:24:48 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:03:41.010 Waiting for target to run... 00:03:41.010 17:24:48 json_config -- json_config/common.sh@25 -- # waitforlisten 4003230 /var/tmp/spdk_tgt.sock 00:03:41.010 17:24:48 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:03:41.010 17:24:48 json_config -- common/autotest_common.sh@831 -- # '[' -z 4003230 ']' 00:03:41.010 17:24:48 json_config -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:03:41.010 17:24:48 json_config -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:41.010 17:24:48 json_config -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:03:41.010 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:03:41.010 17:24:48 json_config -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:41.010 17:24:48 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:41.010 [2024-10-17 17:24:48.758524] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:41.010 [2024-10-17 17:24:48.758596] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4003230 ] 00:03:41.272 [2024-10-17 17:24:49.065211] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:41.272 [2024-10-17 17:24:49.095610] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:41.844 [2024-10-17 17:24:49.596847] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:03:41.844 [2024-10-17 17:24:49.629258] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:03:41.844 17:24:49 json_config -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:41.844 17:24:49 json_config -- common/autotest_common.sh@864 -- # return 0 00:03:41.844 17:24:49 json_config -- json_config/common.sh@26 -- # echo '' 00:03:41.844 00:03:41.844 17:24:49 json_config -- json_config/json_config.sh@380 -- # [[ 0 -eq 1 ]] 00:03:41.844 17:24:49 json_config -- json_config/json_config.sh@384 -- # echo 'INFO: Checking if target configuration is the same...' 00:03:41.844 INFO: Checking if target configuration is the same... 00:03:41.844 17:24:49 json_config -- json_config/json_config.sh@385 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:03:41.844 17:24:49 json_config -- json_config/json_config.sh@385 -- # tgt_rpc save_config 00:03:41.845 17:24:49 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:03:41.845 + '[' 2 -ne 2 ']' 00:03:41.845 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:03:41.845 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:03:41.845 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:03:41.845 +++ basename /dev/fd/62 00:03:41.845 ++ mktemp /tmp/62.XXX 00:03:41.845 + tmp_file_1=/tmp/62.1io 00:03:41.845 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:03:41.845 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:03:41.845 + tmp_file_2=/tmp/spdk_tgt_config.json.fjl 00:03:41.845 + ret=0 00:03:41.845 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:03:42.106 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:03:42.367 + diff -u /tmp/62.1io /tmp/spdk_tgt_config.json.fjl 00:03:42.367 + echo 'INFO: JSON config files are the same' 00:03:42.367 INFO: JSON config files are the same 00:03:42.367 + rm /tmp/62.1io /tmp/spdk_tgt_config.json.fjl 00:03:42.367 + exit 0 00:03:42.367 17:24:50 json_config -- json_config/json_config.sh@386 -- # [[ 0 -eq 1 ]] 00:03:42.367 17:24:50 json_config -- json_config/json_config.sh@391 -- # echo 'INFO: changing configuration and checking if this can be detected...' 00:03:42.367 INFO: changing configuration and checking if this can be detected... 00:03:42.367 17:24:50 json_config -- json_config/json_config.sh@393 -- # tgt_rpc bdev_malloc_delete MallocBdevForConfigChangeCheck 00:03:42.367 17:24:50 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_delete MallocBdevForConfigChangeCheck 00:03:42.367 17:24:50 json_config -- json_config/json_config.sh@394 -- # tgt_rpc save_config 00:03:42.367 17:24:50 json_config -- json_config/json_config.sh@394 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:03:42.367 17:24:50 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:03:42.367 + '[' 2 -ne 2 ']' 00:03:42.367 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:03:42.367 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:03:42.367 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:03:42.367 +++ basename /dev/fd/62 00:03:42.367 ++ mktemp /tmp/62.XXX 00:03:42.367 + tmp_file_1=/tmp/62.W3u 00:03:42.367 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:03:42.367 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:03:42.367 + tmp_file_2=/tmp/spdk_tgt_config.json.U4i 00:03:42.367 + ret=0 00:03:42.367 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:03:42.939 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:03:42.939 + diff -u /tmp/62.W3u /tmp/spdk_tgt_config.json.U4i 00:03:42.939 + ret=1 00:03:42.939 + echo '=== Start of file: /tmp/62.W3u ===' 00:03:42.939 + cat /tmp/62.W3u 00:03:42.939 + echo '=== End of file: /tmp/62.W3u ===' 00:03:42.939 + echo '' 00:03:42.939 + echo '=== Start of file: /tmp/spdk_tgt_config.json.U4i ===' 00:03:42.939 + cat /tmp/spdk_tgt_config.json.U4i 00:03:42.939 + echo '=== End of file: /tmp/spdk_tgt_config.json.U4i ===' 00:03:42.939 + echo '' 00:03:42.939 + rm /tmp/62.W3u /tmp/spdk_tgt_config.json.U4i 00:03:42.939 + exit 1 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@398 -- # echo 'INFO: configuration change detected.' 00:03:42.939 INFO: configuration change detected. 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@401 -- # json_config_test_fini 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@313 -- # timing_enter json_config_test_fini 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@314 -- # local ret=0 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@316 -- # [[ -n '' ]] 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@324 -- # [[ -n 4003230 ]] 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@327 -- # cleanup_bdev_subsystem_config 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@191 -- # timing_enter cleanup_bdev_subsystem_config 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@193 -- # [[ 0 -eq 1 ]] 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@200 -- # uname -s 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@200 -- # [[ Linux = Linux ]] 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@201 -- # rm -f /sample_aio 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@204 -- # [[ 0 -eq 1 ]] 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@208 -- # timing_exit cleanup_bdev_subsystem_config 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:42.939 17:24:50 json_config -- json_config/json_config.sh@330 -- # killprocess 4003230 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@950 -- # '[' -z 4003230 ']' 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@954 -- # kill -0 4003230 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@955 -- # uname 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4003230 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4003230' 00:03:42.939 killing process with pid 4003230 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@969 -- # kill 4003230 00:03:42.939 17:24:50 json_config -- common/autotest_common.sh@974 -- # wait 4003230 00:03:43.200 17:24:51 json_config -- json_config/json_config.sh@333 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:03:43.200 17:24:51 json_config -- json_config/json_config.sh@334 -- # timing_exit json_config_test_fini 00:03:43.200 17:24:51 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:43.200 17:24:51 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:43.200 17:24:51 json_config -- json_config/json_config.sh@335 -- # return 0 00:03:43.200 17:24:51 json_config -- json_config/json_config.sh@403 -- # echo 'INFO: Success' 00:03:43.200 INFO: Success 00:03:43.200 00:03:43.200 real 0m7.285s 00:03:43.200 user 0m8.751s 00:03:43.200 sys 0m1.951s 00:03:43.200 17:24:51 json_config -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:43.200 17:24:51 json_config -- common/autotest_common.sh@10 -- # set +x 00:03:43.200 ************************************ 00:03:43.200 END TEST json_config 00:03:43.200 ************************************ 00:03:43.200 17:24:51 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:03:43.200 17:24:51 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:43.200 17:24:51 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:43.200 17:24:51 -- common/autotest_common.sh@10 -- # set +x 00:03:43.462 ************************************ 00:03:43.462 START TEST json_config_extra_key 00:03:43.462 ************************************ 00:03:43.462 17:24:51 json_config_extra_key -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:03:43.462 17:24:51 json_config_extra_key -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:43.462 17:24:51 json_config_extra_key -- common/autotest_common.sh@1691 -- # lcov --version 00:03:43.462 17:24:51 json_config_extra_key -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:43.462 17:24:51 json_config_extra_key -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:03:43.462 17:24:51 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:43.463 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:43.463 --rc genhtml_branch_coverage=1 00:03:43.463 --rc genhtml_function_coverage=1 00:03:43.463 --rc genhtml_legend=1 00:03:43.463 --rc geninfo_all_blocks=1 00:03:43.463 --rc geninfo_unexecuted_blocks=1 00:03:43.463 00:03:43.463 ' 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:43.463 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:43.463 --rc genhtml_branch_coverage=1 00:03:43.463 --rc genhtml_function_coverage=1 00:03:43.463 --rc genhtml_legend=1 00:03:43.463 --rc geninfo_all_blocks=1 00:03:43.463 --rc geninfo_unexecuted_blocks=1 00:03:43.463 00:03:43.463 ' 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:43.463 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:43.463 --rc genhtml_branch_coverage=1 00:03:43.463 --rc genhtml_function_coverage=1 00:03:43.463 --rc genhtml_legend=1 00:03:43.463 --rc geninfo_all_blocks=1 00:03:43.463 --rc geninfo_unexecuted_blocks=1 00:03:43.463 00:03:43.463 ' 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:43.463 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:43.463 --rc genhtml_branch_coverage=1 00:03:43.463 --rc genhtml_function_coverage=1 00:03:43.463 --rc genhtml_legend=1 00:03:43.463 --rc geninfo_all_blocks=1 00:03:43.463 --rc geninfo_unexecuted_blocks=1 00:03:43.463 00:03:43.463 ' 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:03:43.463 17:24:51 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:03:43.463 17:24:51 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:43.463 17:24:51 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:43.463 17:24:51 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:43.463 17:24:51 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:03:43.463 17:24:51 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:03:43.463 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:03:43.463 17:24:51 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json') 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:03:43.463 INFO: launching applications... 00:03:43.463 17:24:51 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=4003882 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:03:43.463 Waiting for target to run... 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 4003882 /var/tmp/spdk_tgt.sock 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@831 -- # '[' -z 4003882 ']' 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:43.463 17:24:51 json_config_extra_key -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:03:43.463 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:03:43.463 17:24:51 json_config_extra_key -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:43.464 17:24:51 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:03:43.724 [2024-10-17 17:24:51.401227] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:43.724 [2024-10-17 17:24:51.401300] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4003882 ] 00:03:43.984 [2024-10-17 17:24:51.684970] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:43.984 [2024-10-17 17:24:51.711810] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:44.657 17:24:52 json_config_extra_key -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:44.657 17:24:52 json_config_extra_key -- common/autotest_common.sh@864 -- # return 0 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:03:44.657 00:03:44.657 17:24:52 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:03:44.657 INFO: shutting down applications... 00:03:44.657 17:24:52 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 4003882 ]] 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 4003882 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 4003882 00:03:44.657 17:24:52 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:03:44.918 17:24:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:03:44.918 17:24:52 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:03:44.918 17:24:52 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 4003882 00:03:44.918 17:24:52 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:03:44.918 17:24:52 json_config_extra_key -- json_config/common.sh@43 -- # break 00:03:44.918 17:24:52 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:03:44.918 17:24:52 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:03:44.918 SPDK target shutdown done 00:03:44.918 17:24:52 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:03:44.918 Success 00:03:44.918 00:03:44.918 real 0m1.570s 00:03:44.918 user 0m1.183s 00:03:44.918 sys 0m0.408s 00:03:44.918 17:24:52 json_config_extra_key -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:44.918 17:24:52 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:03:44.918 ************************************ 00:03:44.918 END TEST json_config_extra_key 00:03:44.918 ************************************ 00:03:44.918 17:24:52 -- spdk/autotest.sh@161 -- # run_test alias_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:03:44.918 17:24:52 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:44.918 17:24:52 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:44.918 17:24:52 -- common/autotest_common.sh@10 -- # set +x 00:03:44.918 ************************************ 00:03:44.918 START TEST alias_rpc 00:03:44.918 ************************************ 00:03:44.918 17:24:52 alias_rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:03:45.179 * Looking for test storage... 00:03:45.179 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1691 -- # lcov --version 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@345 -- # : 1 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:45.179 17:24:52 alias_rpc -- scripts/common.sh@368 -- # return 0 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:45.179 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:45.179 --rc genhtml_branch_coverage=1 00:03:45.179 --rc genhtml_function_coverage=1 00:03:45.179 --rc genhtml_legend=1 00:03:45.179 --rc geninfo_all_blocks=1 00:03:45.179 --rc geninfo_unexecuted_blocks=1 00:03:45.179 00:03:45.179 ' 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:45.179 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:45.179 --rc genhtml_branch_coverage=1 00:03:45.179 --rc genhtml_function_coverage=1 00:03:45.179 --rc genhtml_legend=1 00:03:45.179 --rc geninfo_all_blocks=1 00:03:45.179 --rc geninfo_unexecuted_blocks=1 00:03:45.179 00:03:45.179 ' 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:45.179 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:45.179 --rc genhtml_branch_coverage=1 00:03:45.179 --rc genhtml_function_coverage=1 00:03:45.179 --rc genhtml_legend=1 00:03:45.179 --rc geninfo_all_blocks=1 00:03:45.179 --rc geninfo_unexecuted_blocks=1 00:03:45.179 00:03:45.179 ' 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:45.179 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:45.179 --rc genhtml_branch_coverage=1 00:03:45.179 --rc genhtml_function_coverage=1 00:03:45.179 --rc genhtml_legend=1 00:03:45.179 --rc geninfo_all_blocks=1 00:03:45.179 --rc geninfo_unexecuted_blocks=1 00:03:45.179 00:03:45.179 ' 00:03:45.179 17:24:52 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:03:45.179 17:24:52 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=4004247 00:03:45.179 17:24:52 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 4004247 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@831 -- # '[' -z 4004247 ']' 00:03:45.179 17:24:52 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:03:45.179 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:45.179 17:24:52 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:45.179 [2024-10-17 17:24:53.039823] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:45.179 [2024-10-17 17:24:53.039901] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4004247 ] 00:03:45.440 [2024-10-17 17:24:53.121754] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:45.440 [2024-10-17 17:24:53.163923] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:46.011 17:24:53 alias_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:46.011 17:24:53 alias_rpc -- common/autotest_common.sh@864 -- # return 0 00:03:46.011 17:24:53 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_config -i 00:03:46.271 17:24:54 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 4004247 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@950 -- # '[' -z 4004247 ']' 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@954 -- # kill -0 4004247 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@955 -- # uname 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4004247 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4004247' 00:03:46.271 killing process with pid 4004247 00:03:46.271 17:24:54 alias_rpc -- common/autotest_common.sh@969 -- # kill 4004247 00:03:46.272 17:24:54 alias_rpc -- common/autotest_common.sh@974 -- # wait 4004247 00:03:46.532 00:03:46.532 real 0m1.520s 00:03:46.532 user 0m1.669s 00:03:46.532 sys 0m0.430s 00:03:46.532 17:24:54 alias_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:46.532 17:24:54 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:03:46.532 ************************************ 00:03:46.532 END TEST alias_rpc 00:03:46.532 ************************************ 00:03:46.532 17:24:54 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:03:46.532 17:24:54 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:03:46.532 17:24:54 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:46.532 17:24:54 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:46.532 17:24:54 -- common/autotest_common.sh@10 -- # set +x 00:03:46.532 ************************************ 00:03:46.532 START TEST spdkcli_tcp 00:03:46.532 ************************************ 00:03:46.532 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:03:46.794 * Looking for test storage... 00:03:46.794 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1691 -- # lcov --version 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:46.794 17:24:54 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:46.794 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:46.794 --rc genhtml_branch_coverage=1 00:03:46.794 --rc genhtml_function_coverage=1 00:03:46.794 --rc genhtml_legend=1 00:03:46.794 --rc geninfo_all_blocks=1 00:03:46.794 --rc geninfo_unexecuted_blocks=1 00:03:46.794 00:03:46.794 ' 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:46.794 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:46.794 --rc genhtml_branch_coverage=1 00:03:46.794 --rc genhtml_function_coverage=1 00:03:46.794 --rc genhtml_legend=1 00:03:46.794 --rc geninfo_all_blocks=1 00:03:46.794 --rc geninfo_unexecuted_blocks=1 00:03:46.794 00:03:46.794 ' 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:46.794 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:46.794 --rc genhtml_branch_coverage=1 00:03:46.794 --rc genhtml_function_coverage=1 00:03:46.794 --rc genhtml_legend=1 00:03:46.794 --rc geninfo_all_blocks=1 00:03:46.794 --rc geninfo_unexecuted_blocks=1 00:03:46.794 00:03:46.794 ' 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:46.794 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:46.794 --rc genhtml_branch_coverage=1 00:03:46.794 --rc genhtml_function_coverage=1 00:03:46.794 --rc genhtml_legend=1 00:03:46.794 --rc geninfo_all_blocks=1 00:03:46.794 --rc geninfo_unexecuted_blocks=1 00:03:46.794 00:03:46.794 ' 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=4004595 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 4004595 00:03:46.794 17:24:54 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@831 -- # '[' -z 4004595 ']' 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:03:46.794 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:46.794 17:24:54 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:03:46.794 [2024-10-17 17:24:54.654423] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:46.795 [2024-10-17 17:24:54.654493] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4004595 ] 00:03:47.055 [2024-10-17 17:24:54.736221] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:03:47.055 [2024-10-17 17:24:54.773571] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:03:47.055 [2024-10-17 17:24:54.773571] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:47.626 17:24:55 spdkcli_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:47.626 17:24:55 spdkcli_tcp -- common/autotest_common.sh@864 -- # return 0 00:03:47.626 17:24:55 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=4004832 00:03:47.626 17:24:55 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:03:47.626 17:24:55 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:03:47.887 [ 00:03:47.887 "bdev_malloc_delete", 00:03:47.887 "bdev_malloc_create", 00:03:47.887 "bdev_null_resize", 00:03:47.887 "bdev_null_delete", 00:03:47.887 "bdev_null_create", 00:03:47.887 "bdev_nvme_cuse_unregister", 00:03:47.887 "bdev_nvme_cuse_register", 00:03:47.887 "bdev_opal_new_user", 00:03:47.887 "bdev_opal_set_lock_state", 00:03:47.887 "bdev_opal_delete", 00:03:47.887 "bdev_opal_get_info", 00:03:47.887 "bdev_opal_create", 00:03:47.887 "bdev_nvme_opal_revert", 00:03:47.887 "bdev_nvme_opal_init", 00:03:47.887 "bdev_nvme_send_cmd", 00:03:47.887 "bdev_nvme_set_keys", 00:03:47.887 "bdev_nvme_get_path_iostat", 00:03:47.887 "bdev_nvme_get_mdns_discovery_info", 00:03:47.887 "bdev_nvme_stop_mdns_discovery", 00:03:47.887 "bdev_nvme_start_mdns_discovery", 00:03:47.887 "bdev_nvme_set_multipath_policy", 00:03:47.887 "bdev_nvme_set_preferred_path", 00:03:47.887 "bdev_nvme_get_io_paths", 00:03:47.887 "bdev_nvme_remove_error_injection", 00:03:47.887 "bdev_nvme_add_error_injection", 00:03:47.887 "bdev_nvme_get_discovery_info", 00:03:47.887 "bdev_nvme_stop_discovery", 00:03:47.887 "bdev_nvme_start_discovery", 00:03:47.887 "bdev_nvme_get_controller_health_info", 00:03:47.887 "bdev_nvme_disable_controller", 00:03:47.887 "bdev_nvme_enable_controller", 00:03:47.887 "bdev_nvme_reset_controller", 00:03:47.887 "bdev_nvme_get_transport_statistics", 00:03:47.887 "bdev_nvme_apply_firmware", 00:03:47.887 "bdev_nvme_detach_controller", 00:03:47.887 "bdev_nvme_get_controllers", 00:03:47.887 "bdev_nvme_attach_controller", 00:03:47.887 "bdev_nvme_set_hotplug", 00:03:47.887 "bdev_nvme_set_options", 00:03:47.888 "bdev_passthru_delete", 00:03:47.888 "bdev_passthru_create", 00:03:47.888 "bdev_lvol_set_parent_bdev", 00:03:47.888 "bdev_lvol_set_parent", 00:03:47.888 "bdev_lvol_check_shallow_copy", 00:03:47.888 "bdev_lvol_start_shallow_copy", 00:03:47.888 "bdev_lvol_grow_lvstore", 00:03:47.888 "bdev_lvol_get_lvols", 00:03:47.888 "bdev_lvol_get_lvstores", 00:03:47.888 "bdev_lvol_delete", 00:03:47.888 "bdev_lvol_set_read_only", 00:03:47.888 "bdev_lvol_resize", 00:03:47.888 "bdev_lvol_decouple_parent", 00:03:47.888 "bdev_lvol_inflate", 00:03:47.888 "bdev_lvol_rename", 00:03:47.888 "bdev_lvol_clone_bdev", 00:03:47.888 "bdev_lvol_clone", 00:03:47.888 "bdev_lvol_snapshot", 00:03:47.888 "bdev_lvol_create", 00:03:47.888 "bdev_lvol_delete_lvstore", 00:03:47.888 "bdev_lvol_rename_lvstore", 00:03:47.888 "bdev_lvol_create_lvstore", 00:03:47.888 "bdev_raid_set_options", 00:03:47.888 "bdev_raid_remove_base_bdev", 00:03:47.888 "bdev_raid_add_base_bdev", 00:03:47.888 "bdev_raid_delete", 00:03:47.888 "bdev_raid_create", 00:03:47.888 "bdev_raid_get_bdevs", 00:03:47.888 "bdev_error_inject_error", 00:03:47.888 "bdev_error_delete", 00:03:47.888 "bdev_error_create", 00:03:47.888 "bdev_split_delete", 00:03:47.888 "bdev_split_create", 00:03:47.888 "bdev_delay_delete", 00:03:47.888 "bdev_delay_create", 00:03:47.888 "bdev_delay_update_latency", 00:03:47.888 "bdev_zone_block_delete", 00:03:47.888 "bdev_zone_block_create", 00:03:47.888 "blobfs_create", 00:03:47.888 "blobfs_detect", 00:03:47.888 "blobfs_set_cache_size", 00:03:47.888 "bdev_aio_delete", 00:03:47.888 "bdev_aio_rescan", 00:03:47.888 "bdev_aio_create", 00:03:47.888 "bdev_ftl_set_property", 00:03:47.888 "bdev_ftl_get_properties", 00:03:47.888 "bdev_ftl_get_stats", 00:03:47.888 "bdev_ftl_unmap", 00:03:47.888 "bdev_ftl_unload", 00:03:47.888 "bdev_ftl_delete", 00:03:47.888 "bdev_ftl_load", 00:03:47.888 "bdev_ftl_create", 00:03:47.888 "bdev_virtio_attach_controller", 00:03:47.888 "bdev_virtio_scsi_get_devices", 00:03:47.888 "bdev_virtio_detach_controller", 00:03:47.888 "bdev_virtio_blk_set_hotplug", 00:03:47.888 "bdev_iscsi_delete", 00:03:47.888 "bdev_iscsi_create", 00:03:47.888 "bdev_iscsi_set_options", 00:03:47.888 "accel_error_inject_error", 00:03:47.888 "ioat_scan_accel_module", 00:03:47.888 "dsa_scan_accel_module", 00:03:47.888 "iaa_scan_accel_module", 00:03:47.888 "vfu_virtio_create_fs_endpoint", 00:03:47.888 "vfu_virtio_create_scsi_endpoint", 00:03:47.888 "vfu_virtio_scsi_remove_target", 00:03:47.888 "vfu_virtio_scsi_add_target", 00:03:47.888 "vfu_virtio_create_blk_endpoint", 00:03:47.888 "vfu_virtio_delete_endpoint", 00:03:47.888 "keyring_file_remove_key", 00:03:47.888 "keyring_file_add_key", 00:03:47.888 "keyring_linux_set_options", 00:03:47.888 "fsdev_aio_delete", 00:03:47.888 "fsdev_aio_create", 00:03:47.888 "iscsi_get_histogram", 00:03:47.888 "iscsi_enable_histogram", 00:03:47.888 "iscsi_set_options", 00:03:47.888 "iscsi_get_auth_groups", 00:03:47.888 "iscsi_auth_group_remove_secret", 00:03:47.888 "iscsi_auth_group_add_secret", 00:03:47.888 "iscsi_delete_auth_group", 00:03:47.888 "iscsi_create_auth_group", 00:03:47.888 "iscsi_set_discovery_auth", 00:03:47.888 "iscsi_get_options", 00:03:47.888 "iscsi_target_node_request_logout", 00:03:47.888 "iscsi_target_node_set_redirect", 00:03:47.888 "iscsi_target_node_set_auth", 00:03:47.888 "iscsi_target_node_add_lun", 00:03:47.888 "iscsi_get_stats", 00:03:47.888 "iscsi_get_connections", 00:03:47.888 "iscsi_portal_group_set_auth", 00:03:47.888 "iscsi_start_portal_group", 00:03:47.888 "iscsi_delete_portal_group", 00:03:47.888 "iscsi_create_portal_group", 00:03:47.888 "iscsi_get_portal_groups", 00:03:47.888 "iscsi_delete_target_node", 00:03:47.888 "iscsi_target_node_remove_pg_ig_maps", 00:03:47.888 "iscsi_target_node_add_pg_ig_maps", 00:03:47.888 "iscsi_create_target_node", 00:03:47.888 "iscsi_get_target_nodes", 00:03:47.888 "iscsi_delete_initiator_group", 00:03:47.888 "iscsi_initiator_group_remove_initiators", 00:03:47.888 "iscsi_initiator_group_add_initiators", 00:03:47.888 "iscsi_create_initiator_group", 00:03:47.888 "iscsi_get_initiator_groups", 00:03:47.888 "nvmf_set_crdt", 00:03:47.888 "nvmf_set_config", 00:03:47.888 "nvmf_set_max_subsystems", 00:03:47.888 "nvmf_stop_mdns_prr", 00:03:47.888 "nvmf_publish_mdns_prr", 00:03:47.888 "nvmf_subsystem_get_listeners", 00:03:47.888 "nvmf_subsystem_get_qpairs", 00:03:47.888 "nvmf_subsystem_get_controllers", 00:03:47.888 "nvmf_get_stats", 00:03:47.888 "nvmf_get_transports", 00:03:47.888 "nvmf_create_transport", 00:03:47.888 "nvmf_get_targets", 00:03:47.888 "nvmf_delete_target", 00:03:47.888 "nvmf_create_target", 00:03:47.888 "nvmf_subsystem_allow_any_host", 00:03:47.888 "nvmf_subsystem_set_keys", 00:03:47.888 "nvmf_subsystem_remove_host", 00:03:47.888 "nvmf_subsystem_add_host", 00:03:47.888 "nvmf_ns_remove_host", 00:03:47.888 "nvmf_ns_add_host", 00:03:47.888 "nvmf_subsystem_remove_ns", 00:03:47.888 "nvmf_subsystem_set_ns_ana_group", 00:03:47.888 "nvmf_subsystem_add_ns", 00:03:47.888 "nvmf_subsystem_listener_set_ana_state", 00:03:47.888 "nvmf_discovery_get_referrals", 00:03:47.888 "nvmf_discovery_remove_referral", 00:03:47.888 "nvmf_discovery_add_referral", 00:03:47.888 "nvmf_subsystem_remove_listener", 00:03:47.888 "nvmf_subsystem_add_listener", 00:03:47.888 "nvmf_delete_subsystem", 00:03:47.888 "nvmf_create_subsystem", 00:03:47.888 "nvmf_get_subsystems", 00:03:47.888 "env_dpdk_get_mem_stats", 00:03:47.888 "nbd_get_disks", 00:03:47.888 "nbd_stop_disk", 00:03:47.888 "nbd_start_disk", 00:03:47.888 "ublk_recover_disk", 00:03:47.888 "ublk_get_disks", 00:03:47.888 "ublk_stop_disk", 00:03:47.888 "ublk_start_disk", 00:03:47.888 "ublk_destroy_target", 00:03:47.888 "ublk_create_target", 00:03:47.888 "virtio_blk_create_transport", 00:03:47.888 "virtio_blk_get_transports", 00:03:47.888 "vhost_controller_set_coalescing", 00:03:47.888 "vhost_get_controllers", 00:03:47.888 "vhost_delete_controller", 00:03:47.888 "vhost_create_blk_controller", 00:03:47.888 "vhost_scsi_controller_remove_target", 00:03:47.888 "vhost_scsi_controller_add_target", 00:03:47.888 "vhost_start_scsi_controller", 00:03:47.888 "vhost_create_scsi_controller", 00:03:47.888 "thread_set_cpumask", 00:03:47.888 "scheduler_set_options", 00:03:47.888 "framework_get_governor", 00:03:47.888 "framework_get_scheduler", 00:03:47.888 "framework_set_scheduler", 00:03:47.888 "framework_get_reactors", 00:03:47.888 "thread_get_io_channels", 00:03:47.888 "thread_get_pollers", 00:03:47.888 "thread_get_stats", 00:03:47.888 "framework_monitor_context_switch", 00:03:47.888 "spdk_kill_instance", 00:03:47.888 "log_enable_timestamps", 00:03:47.888 "log_get_flags", 00:03:47.888 "log_clear_flag", 00:03:47.888 "log_set_flag", 00:03:47.888 "log_get_level", 00:03:47.888 "log_set_level", 00:03:47.888 "log_get_print_level", 00:03:47.888 "log_set_print_level", 00:03:47.888 "framework_enable_cpumask_locks", 00:03:47.888 "framework_disable_cpumask_locks", 00:03:47.888 "framework_wait_init", 00:03:47.888 "framework_start_init", 00:03:47.888 "scsi_get_devices", 00:03:47.888 "bdev_get_histogram", 00:03:47.888 "bdev_enable_histogram", 00:03:47.888 "bdev_set_qos_limit", 00:03:47.888 "bdev_set_qd_sampling_period", 00:03:47.888 "bdev_get_bdevs", 00:03:47.888 "bdev_reset_iostat", 00:03:47.888 "bdev_get_iostat", 00:03:47.888 "bdev_examine", 00:03:47.888 "bdev_wait_for_examine", 00:03:47.888 "bdev_set_options", 00:03:47.888 "accel_get_stats", 00:03:47.888 "accel_set_options", 00:03:47.888 "accel_set_driver", 00:03:47.888 "accel_crypto_key_destroy", 00:03:47.888 "accel_crypto_keys_get", 00:03:47.888 "accel_crypto_key_create", 00:03:47.888 "accel_assign_opc", 00:03:47.888 "accel_get_module_info", 00:03:47.888 "accel_get_opc_assignments", 00:03:47.888 "vmd_rescan", 00:03:47.888 "vmd_remove_device", 00:03:47.888 "vmd_enable", 00:03:47.888 "sock_get_default_impl", 00:03:47.888 "sock_set_default_impl", 00:03:47.888 "sock_impl_set_options", 00:03:47.888 "sock_impl_get_options", 00:03:47.888 "iobuf_get_stats", 00:03:47.888 "iobuf_set_options", 00:03:47.888 "keyring_get_keys", 00:03:47.888 "vfu_tgt_set_base_path", 00:03:47.888 "framework_get_pci_devices", 00:03:47.888 "framework_get_config", 00:03:47.888 "framework_get_subsystems", 00:03:47.888 "fsdev_set_opts", 00:03:47.888 "fsdev_get_opts", 00:03:47.888 "trace_get_info", 00:03:47.888 "trace_get_tpoint_group_mask", 00:03:47.888 "trace_disable_tpoint_group", 00:03:47.888 "trace_enable_tpoint_group", 00:03:47.888 "trace_clear_tpoint_mask", 00:03:47.888 "trace_set_tpoint_mask", 00:03:47.888 "notify_get_notifications", 00:03:47.888 "notify_get_types", 00:03:47.888 "spdk_get_version", 00:03:47.888 "rpc_get_methods" 00:03:47.888 ] 00:03:47.888 17:24:55 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:03:47.888 17:24:55 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:03:47.888 17:24:55 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 4004595 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@950 -- # '[' -z 4004595 ']' 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@954 -- # kill -0 4004595 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@955 -- # uname 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4004595 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4004595' 00:03:47.888 killing process with pid 4004595 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@969 -- # kill 4004595 00:03:47.888 17:24:55 spdkcli_tcp -- common/autotest_common.sh@974 -- # wait 4004595 00:03:48.149 00:03:48.149 real 0m1.519s 00:03:48.149 user 0m2.741s 00:03:48.149 sys 0m0.469s 00:03:48.149 17:24:55 spdkcli_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:48.149 17:24:55 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:03:48.149 ************************************ 00:03:48.149 END TEST spdkcli_tcp 00:03:48.149 ************************************ 00:03:48.149 17:24:55 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:03:48.149 17:24:55 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:48.149 17:24:55 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:48.149 17:24:55 -- common/autotest_common.sh@10 -- # set +x 00:03:48.149 ************************************ 00:03:48.149 START TEST dpdk_mem_utility 00:03:48.149 ************************************ 00:03:48.149 17:24:55 dpdk_mem_utility -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:03:48.149 * Looking for test storage... 00:03:48.149 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility 00:03:48.149 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:48.410 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1691 -- # lcov --version 00:03:48.410 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:48.410 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:48.410 17:24:56 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:03:48.411 17:24:56 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:03:48.411 17:24:56 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:48.411 17:24:56 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:48.411 17:24:56 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:48.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:48.411 --rc genhtml_branch_coverage=1 00:03:48.411 --rc genhtml_function_coverage=1 00:03:48.411 --rc genhtml_legend=1 00:03:48.411 --rc geninfo_all_blocks=1 00:03:48.411 --rc geninfo_unexecuted_blocks=1 00:03:48.411 00:03:48.411 ' 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:48.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:48.411 --rc genhtml_branch_coverage=1 00:03:48.411 --rc genhtml_function_coverage=1 00:03:48.411 --rc genhtml_legend=1 00:03:48.411 --rc geninfo_all_blocks=1 00:03:48.411 --rc geninfo_unexecuted_blocks=1 00:03:48.411 00:03:48.411 ' 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:48.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:48.411 --rc genhtml_branch_coverage=1 00:03:48.411 --rc genhtml_function_coverage=1 00:03:48.411 --rc genhtml_legend=1 00:03:48.411 --rc geninfo_all_blocks=1 00:03:48.411 --rc geninfo_unexecuted_blocks=1 00:03:48.411 00:03:48.411 ' 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:48.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:48.411 --rc genhtml_branch_coverage=1 00:03:48.411 --rc genhtml_function_coverage=1 00:03:48.411 --rc genhtml_legend=1 00:03:48.411 --rc geninfo_all_blocks=1 00:03:48.411 --rc geninfo_unexecuted_blocks=1 00:03:48.411 00:03:48.411 ' 00:03:48.411 17:24:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:03:48.411 17:24:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=4004937 00:03:48.411 17:24:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 4004937 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@831 -- # '[' -z 4004937 ']' 00:03:48.411 17:24:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:03:48.411 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:48.411 17:24:56 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:03:48.411 [2024-10-17 17:24:56.229930] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:48.411 [2024-10-17 17:24:56.230003] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4004937 ] 00:03:48.411 [2024-10-17 17:24:56.311525] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:48.671 [2024-10-17 17:24:56.348012] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:49.294 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:49.294 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@864 -- # return 0 00:03:49.294 17:24:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:03:49.294 17:24:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:03:49.294 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:49.294 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:03:49.294 { 00:03:49.294 "filename": "/tmp/spdk_mem_dump.txt" 00:03:49.294 } 00:03:49.294 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:49.294 17:24:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:03:49.294 DPDK memory size 810.000000 MiB in 1 heap(s) 00:03:49.294 1 heaps totaling size 810.000000 MiB 00:03:49.294 size: 810.000000 MiB heap id: 0 00:03:49.294 end heaps---------- 00:03:49.294 9 mempools totaling size 595.772034 MiB 00:03:49.294 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:03:49.294 size: 158.602051 MiB name: PDU_data_out_Pool 00:03:49.294 size: 92.545471 MiB name: bdev_io_4004937 00:03:49.294 size: 50.003479 MiB name: msgpool_4004937 00:03:49.294 size: 36.509338 MiB name: fsdev_io_4004937 00:03:49.294 size: 21.763794 MiB name: PDU_Pool 00:03:49.294 size: 19.513306 MiB name: SCSI_TASK_Pool 00:03:49.294 size: 4.133484 MiB name: evtpool_4004937 00:03:49.294 size: 0.026123 MiB name: Session_Pool 00:03:49.294 end mempools------- 00:03:49.294 6 memzones totaling size 4.142822 MiB 00:03:49.294 size: 1.000366 MiB name: RG_ring_0_4004937 00:03:49.294 size: 1.000366 MiB name: RG_ring_1_4004937 00:03:49.294 size: 1.000366 MiB name: RG_ring_4_4004937 00:03:49.294 size: 1.000366 MiB name: RG_ring_5_4004937 00:03:49.294 size: 0.125366 MiB name: RG_ring_2_4004937 00:03:49.294 size: 0.015991 MiB name: RG_ring_3_4004937 00:03:49.294 end memzones------- 00:03:49.294 17:24:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py -m 0 00:03:49.294 heap id: 0 total size: 810.000000 MiB number of busy elements: 44 number of free elements: 15 00:03:49.294 list of free elements. size: 10.862488 MiB 00:03:49.294 element at address: 0x200018a00000 with size: 0.999878 MiB 00:03:49.294 element at address: 0x200018c00000 with size: 0.999878 MiB 00:03:49.294 element at address: 0x200000400000 with size: 0.998535 MiB 00:03:49.294 element at address: 0x200031800000 with size: 0.994446 MiB 00:03:49.294 element at address: 0x200006400000 with size: 0.959839 MiB 00:03:49.294 element at address: 0x200012c00000 with size: 0.954285 MiB 00:03:49.294 element at address: 0x200018e00000 with size: 0.936584 MiB 00:03:49.294 element at address: 0x200000200000 with size: 0.717346 MiB 00:03:49.294 element at address: 0x20001a600000 with size: 0.582886 MiB 00:03:49.294 element at address: 0x200000c00000 with size: 0.495422 MiB 00:03:49.294 element at address: 0x20000a600000 with size: 0.490723 MiB 00:03:49.294 element at address: 0x200019000000 with size: 0.485657 MiB 00:03:49.294 element at address: 0x200003e00000 with size: 0.481934 MiB 00:03:49.294 element at address: 0x200027a00000 with size: 0.410034 MiB 00:03:49.294 element at address: 0x200000800000 with size: 0.355042 MiB 00:03:49.294 list of standard malloc elements. size: 199.218628 MiB 00:03:49.294 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:03:49.294 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:03:49.294 element at address: 0x200018afff80 with size: 1.000122 MiB 00:03:49.295 element at address: 0x200018cfff80 with size: 1.000122 MiB 00:03:49.295 element at address: 0x200018efff80 with size: 1.000122 MiB 00:03:49.295 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:03:49.295 element at address: 0x200018eeff00 with size: 0.062622 MiB 00:03:49.295 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:03:49.295 element at address: 0x200018eefdc0 with size: 0.000305 MiB 00:03:49.295 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000004ffb80 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000085ae40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000085b040 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000085f300 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000087f680 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200000cff000 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200003efb980 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200012cf44c0 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200018eefc40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200018eefd00 with size: 0.000183 MiB 00:03:49.295 element at address: 0x2000190bc740 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20001a695380 with size: 0.000183 MiB 00:03:49.295 element at address: 0x20001a695440 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200027a68f80 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200027a69040 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200027a6fc40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200027a6fe40 with size: 0.000183 MiB 00:03:49.295 element at address: 0x200027a6ff00 with size: 0.000183 MiB 00:03:49.295 list of memzone associated elements. size: 599.918884 MiB 00:03:49.295 element at address: 0x20001a695500 with size: 211.416748 MiB 00:03:49.295 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:03:49.295 element at address: 0x200027a6ffc0 with size: 157.562561 MiB 00:03:49.295 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:03:49.295 element at address: 0x200012df4780 with size: 92.045044 MiB 00:03:49.295 associated memzone info: size: 92.044922 MiB name: MP_bdev_io_4004937_0 00:03:49.295 element at address: 0x200000dff380 with size: 48.003052 MiB 00:03:49.295 associated memzone info: size: 48.002930 MiB name: MP_msgpool_4004937_0 00:03:49.295 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:03:49.295 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_4004937_0 00:03:49.295 element at address: 0x2000191be940 with size: 20.255554 MiB 00:03:49.295 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:03:49.295 element at address: 0x2000319feb40 with size: 18.005066 MiB 00:03:49.295 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:03:49.295 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:03:49.295 associated memzone info: size: 3.000122 MiB name: MP_evtpool_4004937_0 00:03:49.295 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:03:49.295 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_4004937 00:03:49.295 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:03:49.295 associated memzone info: size: 1.007996 MiB name: MP_evtpool_4004937 00:03:49.295 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:03:49.295 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:03:49.295 element at address: 0x2000190bc800 with size: 1.008118 MiB 00:03:49.295 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:03:49.295 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:03:49.295 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:03:49.295 element at address: 0x200003efba40 with size: 1.008118 MiB 00:03:49.295 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:03:49.295 element at address: 0x200000cff180 with size: 1.000488 MiB 00:03:49.295 associated memzone info: size: 1.000366 MiB name: RG_ring_0_4004937 00:03:49.295 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:03:49.295 associated memzone info: size: 1.000366 MiB name: RG_ring_1_4004937 00:03:49.295 element at address: 0x200012cf4580 with size: 1.000488 MiB 00:03:49.295 associated memzone info: size: 1.000366 MiB name: RG_ring_4_4004937 00:03:49.295 element at address: 0x2000318fe940 with size: 1.000488 MiB 00:03:49.295 associated memzone info: size: 1.000366 MiB name: RG_ring_5_4004937 00:03:49.295 element at address: 0x20000087f740 with size: 0.500488 MiB 00:03:49.295 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_4004937 00:03:49.295 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:03:49.296 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_4004937 00:03:49.296 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:03:49.296 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:03:49.296 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:03:49.296 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:03:49.296 element at address: 0x20001907c540 with size: 0.250488 MiB 00:03:49.296 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:03:49.296 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:03:49.296 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_4004937 00:03:49.296 element at address: 0x20000085f3c0 with size: 0.125488 MiB 00:03:49.296 associated memzone info: size: 0.125366 MiB name: RG_ring_2_4004937 00:03:49.296 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:03:49.296 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:03:49.296 element at address: 0x200027a69100 with size: 0.023743 MiB 00:03:49.296 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:03:49.296 element at address: 0x20000085b100 with size: 0.016113 MiB 00:03:49.296 associated memzone info: size: 0.015991 MiB name: RG_ring_3_4004937 00:03:49.296 element at address: 0x200027a6f240 with size: 0.002441 MiB 00:03:49.296 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:03:49.296 element at address: 0x2000004ffc40 with size: 0.000305 MiB 00:03:49.296 associated memzone info: size: 0.000183 MiB name: MP_msgpool_4004937 00:03:49.296 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:03:49.296 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_4004937 00:03:49.296 element at address: 0x20000085af00 with size: 0.000305 MiB 00:03:49.296 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_4004937 00:03:49.296 element at address: 0x200027a6fd00 with size: 0.000305 MiB 00:03:49.296 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:03:49.296 17:24:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:03:49.296 17:24:57 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 4004937 00:03:49.296 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@950 -- # '[' -z 4004937 ']' 00:03:49.296 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@954 -- # kill -0 4004937 00:03:49.296 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@955 -- # uname 00:03:49.296 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:49.297 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4004937 00:03:49.577 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:03:49.577 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:03:49.577 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4004937' 00:03:49.577 killing process with pid 4004937 00:03:49.577 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@969 -- # kill 4004937 00:03:49.577 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@974 -- # wait 4004937 00:03:49.577 00:03:49.577 real 0m1.392s 00:03:49.577 user 0m1.450s 00:03:49.577 sys 0m0.418s 00:03:49.577 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:49.577 17:24:57 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:03:49.577 ************************************ 00:03:49.577 END TEST dpdk_mem_utility 00:03:49.577 ************************************ 00:03:49.577 17:24:57 -- spdk/autotest.sh@168 -- # run_test event /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:03:49.577 17:24:57 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:49.577 17:24:57 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:49.577 17:24:57 -- common/autotest_common.sh@10 -- # set +x 00:03:49.577 ************************************ 00:03:49.577 START TEST event 00:03:49.577 ************************************ 00:03:49.577 17:24:57 event -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:03:49.844 * Looking for test storage... 00:03:49.844 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1691 -- # lcov --version 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:49.844 17:24:57 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:49.844 17:24:57 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:49.844 17:24:57 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:49.844 17:24:57 event -- scripts/common.sh@336 -- # IFS=.-: 00:03:49.844 17:24:57 event -- scripts/common.sh@336 -- # read -ra ver1 00:03:49.844 17:24:57 event -- scripts/common.sh@337 -- # IFS=.-: 00:03:49.844 17:24:57 event -- scripts/common.sh@337 -- # read -ra ver2 00:03:49.844 17:24:57 event -- scripts/common.sh@338 -- # local 'op=<' 00:03:49.844 17:24:57 event -- scripts/common.sh@340 -- # ver1_l=2 00:03:49.844 17:24:57 event -- scripts/common.sh@341 -- # ver2_l=1 00:03:49.844 17:24:57 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:49.844 17:24:57 event -- scripts/common.sh@344 -- # case "$op" in 00:03:49.844 17:24:57 event -- scripts/common.sh@345 -- # : 1 00:03:49.844 17:24:57 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:49.844 17:24:57 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:49.844 17:24:57 event -- scripts/common.sh@365 -- # decimal 1 00:03:49.844 17:24:57 event -- scripts/common.sh@353 -- # local d=1 00:03:49.844 17:24:57 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:49.844 17:24:57 event -- scripts/common.sh@355 -- # echo 1 00:03:49.844 17:24:57 event -- scripts/common.sh@365 -- # ver1[v]=1 00:03:49.844 17:24:57 event -- scripts/common.sh@366 -- # decimal 2 00:03:49.844 17:24:57 event -- scripts/common.sh@353 -- # local d=2 00:03:49.844 17:24:57 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:49.844 17:24:57 event -- scripts/common.sh@355 -- # echo 2 00:03:49.844 17:24:57 event -- scripts/common.sh@366 -- # ver2[v]=2 00:03:49.844 17:24:57 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:49.844 17:24:57 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:49.844 17:24:57 event -- scripts/common.sh@368 -- # return 0 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:49.844 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:49.844 --rc genhtml_branch_coverage=1 00:03:49.844 --rc genhtml_function_coverage=1 00:03:49.844 --rc genhtml_legend=1 00:03:49.844 --rc geninfo_all_blocks=1 00:03:49.844 --rc geninfo_unexecuted_blocks=1 00:03:49.844 00:03:49.844 ' 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:49.844 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:49.844 --rc genhtml_branch_coverage=1 00:03:49.844 --rc genhtml_function_coverage=1 00:03:49.844 --rc genhtml_legend=1 00:03:49.844 --rc geninfo_all_blocks=1 00:03:49.844 --rc geninfo_unexecuted_blocks=1 00:03:49.844 00:03:49.844 ' 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:49.844 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:49.844 --rc genhtml_branch_coverage=1 00:03:49.844 --rc genhtml_function_coverage=1 00:03:49.844 --rc genhtml_legend=1 00:03:49.844 --rc geninfo_all_blocks=1 00:03:49.844 --rc geninfo_unexecuted_blocks=1 00:03:49.844 00:03:49.844 ' 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:49.844 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:49.844 --rc genhtml_branch_coverage=1 00:03:49.844 --rc genhtml_function_coverage=1 00:03:49.844 --rc genhtml_legend=1 00:03:49.844 --rc geninfo_all_blocks=1 00:03:49.844 --rc geninfo_unexecuted_blocks=1 00:03:49.844 00:03:49.844 ' 00:03:49.844 17:24:57 event -- event/event.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/nbd_common.sh 00:03:49.844 17:24:57 event -- bdev/nbd_common.sh@6 -- # set -e 00:03:49.844 17:24:57 event -- event/event.sh@45 -- # run_test event_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1101 -- # '[' 6 -le 1 ']' 00:03:49.844 17:24:57 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:49.844 17:24:57 event -- common/autotest_common.sh@10 -- # set +x 00:03:49.844 ************************************ 00:03:49.844 START TEST event_perf 00:03:49.844 ************************************ 00:03:49.844 17:24:57 event.event_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:03:49.844 Running I/O for 1 seconds...[2024-10-17 17:24:57.697178] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:49.844 [2024-10-17 17:24:57.697289] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4005319 ] 00:03:50.105 [2024-10-17 17:24:57.781056] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:03:50.105 [2024-10-17 17:24:57.820809] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:03:50.105 [2024-10-17 17:24:57.821082] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:03:50.105 [2024-10-17 17:24:57.821233] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:50.105 Running I/O for 1 seconds...[2024-10-17 17:24:57.821234] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:03:51.048 00:03:51.048 lcore 0: 185421 00:03:51.048 lcore 1: 185424 00:03:51.048 lcore 2: 185424 00:03:51.048 lcore 3: 185421 00:03:51.048 done. 00:03:51.048 00:03:51.048 real 0m1.175s 00:03:51.048 user 0m4.078s 00:03:51.048 sys 0m0.094s 00:03:51.048 17:24:58 event.event_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:51.048 17:24:58 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:03:51.048 ************************************ 00:03:51.048 END TEST event_perf 00:03:51.048 ************************************ 00:03:51.048 17:24:58 event -- event/event.sh@46 -- # run_test event_reactor /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:03:51.048 17:24:58 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:03:51.048 17:24:58 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:51.048 17:24:58 event -- common/autotest_common.sh@10 -- # set +x 00:03:51.048 ************************************ 00:03:51.048 START TEST event_reactor 00:03:51.048 ************************************ 00:03:51.048 17:24:58 event.event_reactor -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:03:51.048 [2024-10-17 17:24:58.950972] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:51.048 [2024-10-17 17:24:58.951076] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4005670 ] 00:03:51.309 [2024-10-17 17:24:59.035752] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:51.309 [2024-10-17 17:24:59.067846] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:52.251 test_start 00:03:52.251 oneshot 00:03:52.251 tick 100 00:03:52.251 tick 100 00:03:52.251 tick 250 00:03:52.251 tick 100 00:03:52.251 tick 100 00:03:52.251 tick 100 00:03:52.251 tick 250 00:03:52.251 tick 500 00:03:52.251 tick 100 00:03:52.251 tick 100 00:03:52.252 tick 250 00:03:52.252 tick 100 00:03:52.252 tick 100 00:03:52.252 test_end 00:03:52.252 00:03:52.252 real 0m1.168s 00:03:52.252 user 0m1.085s 00:03:52.252 sys 0m0.079s 00:03:52.252 17:25:00 event.event_reactor -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:52.252 17:25:00 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:03:52.252 ************************************ 00:03:52.252 END TEST event_reactor 00:03:52.252 ************************************ 00:03:52.252 17:25:00 event -- event/event.sh@47 -- # run_test event_reactor_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:03:52.252 17:25:00 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:03:52.252 17:25:00 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:52.252 17:25:00 event -- common/autotest_common.sh@10 -- # set +x 00:03:52.513 ************************************ 00:03:52.513 START TEST event_reactor_perf 00:03:52.513 ************************************ 00:03:52.513 17:25:00 event.event_reactor_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:03:52.513 [2024-10-17 17:25:00.198322] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:52.513 [2024-10-17 17:25:00.198427] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4006021 ] 00:03:52.513 [2024-10-17 17:25:00.279075] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:03:52.513 [2024-10-17 17:25:00.311174] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:53.454 test_start 00:03:53.454 test_end 00:03:53.454 Performance: 537681 events per second 00:03:53.454 00:03:53.454 real 0m1.160s 00:03:53.454 user 0m1.080s 00:03:53.454 sys 0m0.078s 00:03:53.454 17:25:01 event.event_reactor_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:53.454 17:25:01 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:03:53.454 ************************************ 00:03:53.454 END TEST event_reactor_perf 00:03:53.454 ************************************ 00:03:53.715 17:25:01 event -- event/event.sh@49 -- # uname -s 00:03:53.715 17:25:01 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:03:53.715 17:25:01 event -- event/event.sh@50 -- # run_test event_scheduler /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:03:53.715 17:25:01 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:53.715 17:25:01 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:53.715 17:25:01 event -- common/autotest_common.sh@10 -- # set +x 00:03:53.715 ************************************ 00:03:53.715 START TEST event_scheduler 00:03:53.715 ************************************ 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:03:53.715 * Looking for test storage... 00:03:53.715 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1691 -- # lcov --version 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:03:53.715 17:25:01 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:03:53.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:53.715 --rc genhtml_branch_coverage=1 00:03:53.715 --rc genhtml_function_coverage=1 00:03:53.715 --rc genhtml_legend=1 00:03:53.715 --rc geninfo_all_blocks=1 00:03:53.715 --rc geninfo_unexecuted_blocks=1 00:03:53.715 00:03:53.715 ' 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:03:53.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:53.715 --rc genhtml_branch_coverage=1 00:03:53.715 --rc genhtml_function_coverage=1 00:03:53.715 --rc genhtml_legend=1 00:03:53.715 --rc geninfo_all_blocks=1 00:03:53.715 --rc geninfo_unexecuted_blocks=1 00:03:53.715 00:03:53.715 ' 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:03:53.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:53.715 --rc genhtml_branch_coverage=1 00:03:53.715 --rc genhtml_function_coverage=1 00:03:53.715 --rc genhtml_legend=1 00:03:53.715 --rc geninfo_all_blocks=1 00:03:53.715 --rc geninfo_unexecuted_blocks=1 00:03:53.715 00:03:53.715 ' 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:03:53.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:03:53.715 --rc genhtml_branch_coverage=1 00:03:53.715 --rc genhtml_function_coverage=1 00:03:53.715 --rc genhtml_legend=1 00:03:53.715 --rc geninfo_all_blocks=1 00:03:53.715 --rc geninfo_unexecuted_blocks=1 00:03:53.715 00:03:53.715 ' 00:03:53.715 17:25:01 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:03:53.715 17:25:01 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=4006329 00:03:53.715 17:25:01 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:03:53.715 17:25:01 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 4006329 00:03:53.715 17:25:01 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@831 -- # '[' -z 4006329 ']' 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:03:53.715 17:25:01 event.event_scheduler -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:53.716 17:25:01 event.event_scheduler -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:03:53.716 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:03:53.716 17:25:01 event.event_scheduler -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:53.716 17:25:01 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:03:53.976 [2024-10-17 17:25:01.687658] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:53.976 [2024-10-17 17:25:01.687740] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4006329 ] 00:03:53.977 [2024-10-17 17:25:01.771196] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:03:53.977 [2024-10-17 17:25:01.827601] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:53.977 [2024-10-17 17:25:01.827773] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:03:53.977 [2024-10-17 17:25:01.827850] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:03:53.977 [2024-10-17 17:25:01.827852] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@864 -- # return 0 00:03:54.920 17:25:02 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 [2024-10-17 17:25:02.490430] dpdk_governor.c: 173:_init: *ERROR*: App core mask contains some but not all of a set of SMT siblings 00:03:54.920 [2024-10-17 17:25:02.490450] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:03:54.920 [2024-10-17 17:25:02.490460] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:03:54.920 [2024-10-17 17:25:02.490466] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:03:54.920 [2024-10-17 17:25:02.490472] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 [2024-10-17 17:25:02.556927] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 ************************************ 00:03:54.920 START TEST scheduler_create_thread 00:03:54.920 ************************************ 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1125 -- # scheduler_create_thread 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 2 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 3 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 4 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 5 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 6 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 7 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 8 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:54.920 9 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:54.920 17:25:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:55.491 10 00:03:55.492 17:25:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:55.492 17:25:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:03:55.492 17:25:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:55.492 17:25:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:56.875 17:25:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:56.875 17:25:04 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:03:56.875 17:25:04 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:03:56.875 17:25:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:56.875 17:25:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:57.446 17:25:05 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:57.446 17:25:05 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:03:57.446 17:25:05 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:57.446 17:25:05 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:58.387 17:25:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:58.387 17:25:06 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:03:58.387 17:25:06 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:03:58.387 17:25:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:03:58.387 17:25:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:58.958 17:25:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:03:58.958 00:03:58.958 real 0m4.225s 00:03:58.958 user 0m0.023s 00:03:58.958 sys 0m0.008s 00:03:58.958 17:25:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:58.959 17:25:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:03:58.959 ************************************ 00:03:58.959 END TEST scheduler_create_thread 00:03:58.959 ************************************ 00:03:58.959 17:25:06 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:03:58.959 17:25:06 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 4006329 00:03:58.959 17:25:06 event.event_scheduler -- common/autotest_common.sh@950 -- # '[' -z 4006329 ']' 00:03:58.959 17:25:06 event.event_scheduler -- common/autotest_common.sh@954 -- # kill -0 4006329 00:03:58.959 17:25:06 event.event_scheduler -- common/autotest_common.sh@955 -- # uname 00:03:58.959 17:25:06 event.event_scheduler -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:03:58.959 17:25:06 event.event_scheduler -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4006329 00:03:59.218 17:25:06 event.event_scheduler -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:03:59.218 17:25:06 event.event_scheduler -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:03:59.218 17:25:06 event.event_scheduler -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4006329' 00:03:59.218 killing process with pid 4006329 00:03:59.218 17:25:06 event.event_scheduler -- common/autotest_common.sh@969 -- # kill 4006329 00:03:59.218 17:25:06 event.event_scheduler -- common/autotest_common.sh@974 -- # wait 4006329 00:03:59.218 [2024-10-17 17:25:07.102522] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:03:59.478 00:03:59.478 real 0m5.838s 00:03:59.478 user 0m12.871s 00:03:59.478 sys 0m0.427s 00:03:59.478 17:25:07 event.event_scheduler -- common/autotest_common.sh@1126 -- # xtrace_disable 00:03:59.478 17:25:07 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:03:59.478 ************************************ 00:03:59.478 END TEST event_scheduler 00:03:59.478 ************************************ 00:03:59.478 17:25:07 event -- event/event.sh@51 -- # modprobe -n nbd 00:03:59.478 17:25:07 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:03:59.478 17:25:07 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:03:59.478 17:25:07 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:03:59.478 17:25:07 event -- common/autotest_common.sh@10 -- # set +x 00:03:59.478 ************************************ 00:03:59.478 START TEST app_repeat 00:03:59.478 ************************************ 00:03:59.478 17:25:07 event.app_repeat -- common/autotest_common.sh@1125 -- # app_repeat_test 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@19 -- # repeat_pid=4007489 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 4007489' 00:03:59.478 Process app_repeat pid: 4007489 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:03:59.478 spdk_app_start Round 0 00:03:59.478 17:25:07 event.app_repeat -- event/event.sh@25 -- # waitforlisten 4007489 /var/tmp/spdk-nbd.sock 00:03:59.478 17:25:07 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 4007489 ']' 00:03:59.478 17:25:07 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:03:59.478 17:25:07 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:03:59.478 17:25:07 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:03:59.478 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:03:59.478 17:25:07 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:03:59.478 17:25:07 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:03:59.478 [2024-10-17 17:25:07.382848] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:03:59.478 [2024-10-17 17:25:07.382909] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4007489 ] 00:03:59.738 [2024-10-17 17:25:07.462043] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:03:59.738 [2024-10-17 17:25:07.493574] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:03:59.738 [2024-10-17 17:25:07.493575] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:03:59.738 17:25:07 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:03:59.738 17:25:07 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:03:59.738 17:25:07 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:03:59.999 Malloc0 00:03:59.999 17:25:07 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:00.261 Malloc1 00:04:00.261 17:25:07 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:00.262 17:25:07 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:04:00.262 /dev/nbd0 00:04:00.262 17:25:08 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:04:00.262 17:25:08 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:00.262 1+0 records in 00:04:00.262 1+0 records out 00:04:00.262 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000295093 s, 13.9 MB/s 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:04:00.262 17:25:08 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:00.523 17:25:08 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:04:00.523 17:25:08 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:04:00.524 /dev/nbd1 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:00.524 1+0 records in 00:04:00.524 1+0 records out 00:04:00.524 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000276051 s, 14.8 MB/s 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:04:00.524 17:25:08 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:00.524 17:25:08 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:00.785 17:25:08 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:04:00.785 { 00:04:00.785 "nbd_device": "/dev/nbd0", 00:04:00.785 "bdev_name": "Malloc0" 00:04:00.785 }, 00:04:00.785 { 00:04:00.785 "nbd_device": "/dev/nbd1", 00:04:00.785 "bdev_name": "Malloc1" 00:04:00.785 } 00:04:00.785 ]' 00:04:00.785 17:25:08 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:04:00.785 { 00:04:00.785 "nbd_device": "/dev/nbd0", 00:04:00.785 "bdev_name": "Malloc0" 00:04:00.785 }, 00:04:00.785 { 00:04:00.785 "nbd_device": "/dev/nbd1", 00:04:00.785 "bdev_name": "Malloc1" 00:04:00.785 } 00:04:00.785 ]' 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:04:00.786 /dev/nbd1' 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:04:00.786 /dev/nbd1' 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:04:00.786 256+0 records in 00:04:00.786 256+0 records out 00:04:00.786 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0120853 s, 86.8 MB/s 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:04:00.786 256+0 records in 00:04:00.786 256+0 records out 00:04:00.786 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.012252 s, 85.6 MB/s 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:00.786 17:25:08 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:04:01.047 256+0 records in 00:04:01.047 256+0 records out 00:04:01.047 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0131096 s, 80.0 MB/s 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:01.047 17:25:08 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:01.308 17:25:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:04:01.577 17:25:09 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:04:01.578 17:25:09 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:04:01.578 17:25:09 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:04:01.844 17:25:09 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:04:01.844 [2024-10-17 17:25:09.633040] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:01.844 [2024-10-17 17:25:09.661642] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:01.844 [2024-10-17 17:25:09.661642] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:01.844 [2024-10-17 17:25:09.690803] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:04:01.844 [2024-10-17 17:25:09.690835] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:04:05.143 17:25:12 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:04:05.143 17:25:12 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:04:05.143 spdk_app_start Round 1 00:04:05.143 17:25:12 event.app_repeat -- event/event.sh@25 -- # waitforlisten 4007489 /var/tmp/spdk-nbd.sock 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 4007489 ']' 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:04:05.143 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:05.143 17:25:12 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:04:05.143 17:25:12 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:05.143 Malloc0 00:04:05.143 17:25:12 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:05.403 Malloc1 00:04:05.404 17:25:13 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:04:05.404 /dev/nbd0 00:04:05.404 17:25:13 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:04:05.663 17:25:13 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:04:05.663 17:25:13 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:04:05.663 17:25:13 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:04:05.663 17:25:13 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:04:05.663 17:25:13 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:04:05.663 17:25:13 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:04:05.663 17:25:13 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:04:05.663 17:25:13 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:05.664 1+0 records in 00:04:05.664 1+0 records out 00:04:05.664 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000282254 s, 14.5 MB/s 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:04:05.664 17:25:13 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:05.664 17:25:13 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:05.664 17:25:13 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:04:05.664 /dev/nbd1 00:04:05.664 17:25:13 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:04:05.664 17:25:13 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:05.664 1+0 records in 00:04:05.664 1+0 records out 00:04:05.664 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000303802 s, 13.5 MB/s 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:04:05.664 17:25:13 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:05.924 17:25:13 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:04:05.924 17:25:13 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:04:05.924 { 00:04:05.924 "nbd_device": "/dev/nbd0", 00:04:05.924 "bdev_name": "Malloc0" 00:04:05.924 }, 00:04:05.924 { 00:04:05.924 "nbd_device": "/dev/nbd1", 00:04:05.924 "bdev_name": "Malloc1" 00:04:05.924 } 00:04:05.924 ]' 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:04:05.924 { 00:04:05.924 "nbd_device": "/dev/nbd0", 00:04:05.924 "bdev_name": "Malloc0" 00:04:05.924 }, 00:04:05.924 { 00:04:05.924 "nbd_device": "/dev/nbd1", 00:04:05.924 "bdev_name": "Malloc1" 00:04:05.924 } 00:04:05.924 ]' 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:04:05.924 /dev/nbd1' 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:04:05.924 /dev/nbd1' 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:04:05.924 256+0 records in 00:04:05.924 256+0 records out 00:04:05.924 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0127371 s, 82.3 MB/s 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:05.924 17:25:13 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:04:06.185 256+0 records in 00:04:06.185 256+0 records out 00:04:06.185 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0122884 s, 85.3 MB/s 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:04:06.185 256+0 records in 00:04:06.185 256+0 records out 00:04:06.185 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0130282 s, 80.5 MB/s 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:06.185 17:25:13 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:06.185 17:25:14 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:06.446 17:25:14 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:04:06.707 17:25:14 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:04:06.707 17:25:14 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:04:06.968 17:25:14 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:04:06.968 [2024-10-17 17:25:14.785970] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:06.968 [2024-10-17 17:25:14.815805] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:06.968 [2024-10-17 17:25:14.815988] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:06.968 [2024-10-17 17:25:14.845708] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:04:06.968 [2024-10-17 17:25:14.845746] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:04:10.271 17:25:17 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:04:10.271 17:25:17 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:04:10.271 spdk_app_start Round 2 00:04:10.271 17:25:17 event.app_repeat -- event/event.sh@25 -- # waitforlisten 4007489 /var/tmp/spdk-nbd.sock 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 4007489 ']' 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:04:10.271 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:10.271 17:25:17 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:04:10.271 17:25:17 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:10.271 Malloc0 00:04:10.271 17:25:18 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:04:10.532 Malloc1 00:04:10.532 17:25:18 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:10.532 17:25:18 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:04:10.793 /dev/nbd0 00:04:10.793 17:25:18 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:04:10.793 17:25:18 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:10.793 1+0 records in 00:04:10.793 1+0 records out 00:04:10.793 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000302238 s, 13.6 MB/s 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:04:10.793 17:25:18 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:04:10.793 17:25:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:10.793 17:25:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:10.793 17:25:18 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:04:10.793 /dev/nbd1 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:04:11.053 1+0 records in 00:04:11.053 1+0 records out 00:04:11.053 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000283158 s, 14.5 MB/s 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:04:11.053 17:25:18 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:04:11.053 { 00:04:11.053 "nbd_device": "/dev/nbd0", 00:04:11.053 "bdev_name": "Malloc0" 00:04:11.053 }, 00:04:11.053 { 00:04:11.053 "nbd_device": "/dev/nbd1", 00:04:11.053 "bdev_name": "Malloc1" 00:04:11.053 } 00:04:11.053 ]' 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:04:11.053 { 00:04:11.053 "nbd_device": "/dev/nbd0", 00:04:11.053 "bdev_name": "Malloc0" 00:04:11.053 }, 00:04:11.053 { 00:04:11.053 "nbd_device": "/dev/nbd1", 00:04:11.053 "bdev_name": "Malloc1" 00:04:11.053 } 00:04:11.053 ]' 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:11.053 17:25:18 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:04:11.053 /dev/nbd1' 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:04:11.314 /dev/nbd1' 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:04:11.314 256+0 records in 00:04:11.314 256+0 records out 00:04:11.314 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0119395 s, 87.8 MB/s 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:11.314 17:25:18 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:04:11.314 256+0 records in 00:04:11.314 256+0 records out 00:04:11.314 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0122514 s, 85.6 MB/s 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:04:11.314 256+0 records in 00:04:11.314 256+0 records out 00:04:11.314 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0130101 s, 80.6 MB/s 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:04:11.314 17:25:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:04:11.315 17:25:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:04:11.574 17:25:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:04:11.575 17:25:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:04:11.575 17:25:19 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:04:11.575 17:25:19 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:04:11.575 17:25:19 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:04:11.575 17:25:19 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:04:11.575 17:25:19 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:04:11.575 17:25:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:04:11.835 17:25:19 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:04:11.835 17:25:19 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:04:12.095 17:25:19 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:04:12.095 [2024-10-17 17:25:19.940797] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:12.095 [2024-10-17 17:25:19.970208] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:12.095 [2024-10-17 17:25:19.970208] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:12.095 [2024-10-17 17:25:19.999439] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:04:12.095 [2024-10-17 17:25:19.999471] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:04:15.392 17:25:22 event.app_repeat -- event/event.sh@38 -- # waitforlisten 4007489 /var/tmp/spdk-nbd.sock 00:04:15.392 17:25:22 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 4007489 ']' 00:04:15.392 17:25:22 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:04:15.392 17:25:22 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:15.392 17:25:22 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:04:15.392 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:04:15.392 17:25:22 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:15.392 17:25:22 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:04:15.392 17:25:23 event.app_repeat -- event/event.sh@39 -- # killprocess 4007489 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@950 -- # '[' -z 4007489 ']' 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@954 -- # kill -0 4007489 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@955 -- # uname 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4007489 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4007489' 00:04:15.392 killing process with pid 4007489 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@969 -- # kill 4007489 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@974 -- # wait 4007489 00:04:15.392 spdk_app_start is called in Round 0. 00:04:15.392 Shutdown signal received, stop current app iteration 00:04:15.392 Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 reinitialization... 00:04:15.392 spdk_app_start is called in Round 1. 00:04:15.392 Shutdown signal received, stop current app iteration 00:04:15.392 Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 reinitialization... 00:04:15.392 spdk_app_start is called in Round 2. 00:04:15.392 Shutdown signal received, stop current app iteration 00:04:15.392 Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 reinitialization... 00:04:15.392 spdk_app_start is called in Round 3. 00:04:15.392 Shutdown signal received, stop current app iteration 00:04:15.392 17:25:23 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:04:15.392 17:25:23 event.app_repeat -- event/event.sh@42 -- # return 0 00:04:15.392 00:04:15.392 real 0m15.854s 00:04:15.392 user 0m34.880s 00:04:15.392 sys 0m2.259s 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:15.392 17:25:23 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:04:15.392 ************************************ 00:04:15.392 END TEST app_repeat 00:04:15.392 ************************************ 00:04:15.392 17:25:23 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:04:15.392 17:25:23 event -- event/event.sh@55 -- # run_test cpu_locks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:04:15.392 17:25:23 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:15.392 17:25:23 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:15.392 17:25:23 event -- common/autotest_common.sh@10 -- # set +x 00:04:15.392 ************************************ 00:04:15.392 START TEST cpu_locks 00:04:15.392 ************************************ 00:04:15.392 17:25:23 event.cpu_locks -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:04:15.653 * Looking for test storage... 00:04:15.653 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1691 -- # lcov --version 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:15.653 17:25:23 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:15.653 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:15.653 --rc genhtml_branch_coverage=1 00:04:15.653 --rc genhtml_function_coverage=1 00:04:15.653 --rc genhtml_legend=1 00:04:15.653 --rc geninfo_all_blocks=1 00:04:15.653 --rc geninfo_unexecuted_blocks=1 00:04:15.653 00:04:15.653 ' 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:15.653 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:15.653 --rc genhtml_branch_coverage=1 00:04:15.653 --rc genhtml_function_coverage=1 00:04:15.653 --rc genhtml_legend=1 00:04:15.653 --rc geninfo_all_blocks=1 00:04:15.653 --rc geninfo_unexecuted_blocks=1 00:04:15.653 00:04:15.653 ' 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:15.653 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:15.653 --rc genhtml_branch_coverage=1 00:04:15.653 --rc genhtml_function_coverage=1 00:04:15.653 --rc genhtml_legend=1 00:04:15.653 --rc geninfo_all_blocks=1 00:04:15.653 --rc geninfo_unexecuted_blocks=1 00:04:15.653 00:04:15.653 ' 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:15.653 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:15.653 --rc genhtml_branch_coverage=1 00:04:15.653 --rc genhtml_function_coverage=1 00:04:15.653 --rc genhtml_legend=1 00:04:15.653 --rc geninfo_all_blocks=1 00:04:15.653 --rc geninfo_unexecuted_blocks=1 00:04:15.653 00:04:15.653 ' 00:04:15.653 17:25:23 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:04:15.653 17:25:23 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:04:15.653 17:25:23 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:04:15.653 17:25:23 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:15.653 17:25:23 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:15.653 ************************************ 00:04:15.653 START TEST default_locks 00:04:15.654 ************************************ 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@1125 -- # default_locks 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=4010875 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 4010875 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 4010875 ']' 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:15.654 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:15.654 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:04:15.915 [2024-10-17 17:25:23.577946] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:15.915 [2024-10-17 17:25:23.578002] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4010875 ] 00:04:15.915 [2024-10-17 17:25:23.635659] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:15.915 [2024-10-17 17:25:23.667786] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:16.175 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:16.175 17:25:23 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 0 00:04:16.175 17:25:23 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 4010875 00:04:16.175 17:25:23 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 4010875 00:04:16.175 17:25:23 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:04:16.435 lslocks: write error 00:04:16.435 17:25:24 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 4010875 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@950 -- # '[' -z 4010875 ']' 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # kill -0 4010875 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # uname 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4010875 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4010875' 00:04:16.436 killing process with pid 4010875 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@969 -- # kill 4010875 00:04:16.436 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@974 -- # wait 4010875 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 4010875 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # local es=0 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 4010875 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # waitforlisten 4010875 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 4010875 ']' 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:16.696 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:16.696 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:04:16.697 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 846: kill: (4010875) - No such process 00:04:16.697 ERROR: process (pid: 4010875) is no longer running 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 1 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # es=1 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:04:16.697 00:04:16.697 real 0m0.952s 00:04:16.697 user 0m0.978s 00:04:16.697 sys 0m0.472s 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:16.697 17:25:24 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:04:16.697 ************************************ 00:04:16.697 END TEST default_locks 00:04:16.697 ************************************ 00:04:16.697 17:25:24 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:04:16.697 17:25:24 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:16.697 17:25:24 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:16.697 17:25:24 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:16.697 ************************************ 00:04:16.697 START TEST default_locks_via_rpc 00:04:16.697 ************************************ 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1125 -- # default_locks_via_rpc 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=4011120 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 4011120 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 4011120 ']' 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:16.697 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:16.697 17:25:24 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:16.697 [2024-10-17 17:25:24.599322] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:16.697 [2024-10-17 17:25:24.599380] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4011120 ] 00:04:16.957 [2024-10-17 17:25:24.682324] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:16.957 [2024-10-17 17:25:24.716013] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:04:17.527 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:04:17.528 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:04:17.528 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:17.528 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:17.528 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:17.528 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 4011120 00:04:17.528 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 4011120 00:04:17.528 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 4011120 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@950 -- # '[' -z 4011120 ']' 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # kill -0 4011120 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # uname 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4011120 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4011120' 00:04:18.206 killing process with pid 4011120 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@969 -- # kill 4011120 00:04:18.206 17:25:25 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@974 -- # wait 4011120 00:04:18.483 00:04:18.483 real 0m1.633s 00:04:18.483 user 0m1.737s 00:04:18.483 sys 0m0.587s 00:04:18.483 17:25:26 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:18.483 17:25:26 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:18.483 ************************************ 00:04:18.483 END TEST default_locks_via_rpc 00:04:18.483 ************************************ 00:04:18.483 17:25:26 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:04:18.484 17:25:26 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:18.484 17:25:26 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:18.484 17:25:26 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:18.484 ************************************ 00:04:18.484 START TEST non_locking_app_on_locked_coremask 00:04:18.484 ************************************ 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # non_locking_app_on_locked_coremask 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=4011493 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 4011493 /var/tmp/spdk.sock 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 4011493 ']' 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:18.484 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:18.484 17:25:26 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:18.484 [2024-10-17 17:25:26.318165] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:18.484 [2024-10-17 17:25:26.318233] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4011493 ] 00:04:18.484 [2024-10-17 17:25:26.398777] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:18.745 [2024-10-17 17:25:26.434734] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=4011701 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 4011701 /var/tmp/spdk2.sock 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 4011701 ']' 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:04:19.318 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:19.318 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:19.318 [2024-10-17 17:25:27.131212] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:19.318 [2024-10-17 17:25:27.131262] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4011701 ] 00:04:19.318 [2024-10-17 17:25:27.202102] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:04:19.318 [2024-10-17 17:25:27.202127] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:19.580 [2024-10-17 17:25:27.264541] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:20.152 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:20.152 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:04:20.152 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 4011493 00:04:20.152 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:04:20.152 17:25:27 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 4011493 00:04:20.723 lslocks: write error 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 4011493 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 4011493 ']' 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 4011493 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4011493 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4011493' 00:04:20.723 killing process with pid 4011493 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 4011493 00:04:20.723 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 4011493 00:04:21.295 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 4011701 00:04:21.295 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 4011701 ']' 00:04:21.295 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 4011701 00:04:21.295 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:04:21.295 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:21.295 17:25:28 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4011701 00:04:21.295 17:25:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:21.295 17:25:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:21.295 17:25:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4011701' 00:04:21.295 killing process with pid 4011701 00:04:21.295 17:25:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 4011701 00:04:21.295 17:25:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 4011701 00:04:21.557 00:04:21.557 real 0m2.967s 00:04:21.557 user 0m3.311s 00:04:21.557 sys 0m0.903s 00:04:21.557 17:25:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:21.557 17:25:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:21.557 ************************************ 00:04:21.557 END TEST non_locking_app_on_locked_coremask 00:04:21.557 ************************************ 00:04:21.557 17:25:29 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:04:21.557 17:25:29 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:21.557 17:25:29 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:21.557 17:25:29 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:21.557 ************************************ 00:04:21.557 START TEST locking_app_on_unlocked_coremask 00:04:21.557 ************************************ 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_unlocked_coremask 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=4012195 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 4012195 /var/tmp/spdk.sock 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 4012195 ']' 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:21.557 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:21.557 17:25:29 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:21.557 [2024-10-17 17:25:29.350142] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:21.557 [2024-10-17 17:25:29.350196] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4012195 ] 00:04:21.557 [2024-10-17 17:25:29.428257] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:04:21.557 [2024-10-17 17:25:29.428284] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:21.557 [2024-10-17 17:25:29.460086] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=4012237 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 4012237 /var/tmp/spdk2.sock 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 4012237 ']' 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:04:22.500 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:22.500 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:22.500 [2024-10-17 17:25:30.195478] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:22.500 [2024-10-17 17:25:30.195533] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4012237 ] 00:04:22.500 [2024-10-17 17:25:30.266925] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:22.500 [2024-10-17 17:25:30.332268] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:23.072 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:23.072 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:04:23.072 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 4012237 00:04:23.072 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 4012237 00:04:23.072 17:25:30 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:04:24.015 lslocks: write error 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 4012195 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 4012195 ']' 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 4012195 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4012195 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4012195' 00:04:24.015 killing process with pid 4012195 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 4012195 00:04:24.015 17:25:31 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 4012195 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 4012237 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 4012237 ']' 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 4012237 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4012237 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4012237' 00:04:24.276 killing process with pid 4012237 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 4012237 00:04:24.276 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 4012237 00:04:24.538 00:04:24.538 real 0m3.005s 00:04:24.538 user 0m3.358s 00:04:24.538 sys 0m0.915s 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:24.538 ************************************ 00:04:24.538 END TEST locking_app_on_unlocked_coremask 00:04:24.538 ************************************ 00:04:24.538 17:25:32 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:04:24.538 17:25:32 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:24.538 17:25:32 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:24.538 17:25:32 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:24.538 ************************************ 00:04:24.538 START TEST locking_app_on_locked_coremask 00:04:24.538 ************************************ 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_locked_coremask 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=4012881 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 4012881 /var/tmp/spdk.sock 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 4012881 ']' 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:24.538 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:24.538 17:25:32 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:24.538 [2024-10-17 17:25:32.431558] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:24.538 [2024-10-17 17:25:32.431614] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4012881 ] 00:04:24.799 [2024-10-17 17:25:32.513025] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:24.799 [2024-10-17 17:25:32.546159] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=4012920 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 4012920 /var/tmp/spdk2.sock 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # local es=0 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 4012920 /var/tmp/spdk2.sock 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # waitforlisten 4012920 /var/tmp/spdk2.sock 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 4012920 ']' 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:04:25.369 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:25.370 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:04:25.370 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:04:25.370 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:25.370 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:25.370 [2024-10-17 17:25:33.256091] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:25.370 [2024-10-17 17:25:33.256153] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4012920 ] 00:04:25.631 [2024-10-17 17:25:33.330001] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 4012881 has claimed it. 00:04:25.631 [2024-10-17 17:25:33.330033] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:04:26.205 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 846: kill: (4012920) - No such process 00:04:26.205 ERROR: process (pid: 4012920) is no longer running 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 1 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # es=1 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 4012881 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 4012881 00:04:26.205 17:25:33 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:04:26.777 lslocks: write error 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 4012881 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 4012881 ']' 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 4012881 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4012881 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4012881' 00:04:26.777 killing process with pid 4012881 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 4012881 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 4012881 00:04:26.777 00:04:26.777 real 0m2.266s 00:04:26.777 user 0m2.538s 00:04:26.777 sys 0m0.654s 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:26.777 17:25:34 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:26.777 ************************************ 00:04:26.777 END TEST locking_app_on_locked_coremask 00:04:26.777 ************************************ 00:04:26.777 17:25:34 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:04:26.777 17:25:34 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:26.777 17:25:34 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:26.777 17:25:34 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:27.038 ************************************ 00:04:27.038 START TEST locking_overlapped_coremask 00:04:27.038 ************************************ 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=4013291 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 4013291 /var/tmp/spdk.sock 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 4013291 ']' 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:27.038 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:27.038 17:25:34 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:27.038 [2024-10-17 17:25:34.769786] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:27.038 [2024-10-17 17:25:34.769842] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4013291 ] 00:04:27.038 [2024-10-17 17:25:34.850952] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:04:27.038 [2024-10-17 17:25:34.886225] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:27.038 [2024-10-17 17:25:34.886375] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:27.038 [2024-10-17 17:25:34.886377] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 0 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=4013484 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 4013484 /var/tmp/spdk2.sock 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # local es=0 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 4013484 /var/tmp/spdk2.sock 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # waitforlisten 4013484 /var/tmp/spdk2.sock 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 4013484 ']' 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:04:27.981 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:27.981 17:25:35 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:27.981 [2024-10-17 17:25:35.629331] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:27.981 [2024-10-17 17:25:35.629385] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4013484 ] 00:04:27.981 [2024-10-17 17:25:35.724195] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 4013291 has claimed it. 00:04:27.981 [2024-10-17 17:25:35.724237] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:04:28.554 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 846: kill: (4013484) - No such process 00:04:28.554 ERROR: process (pid: 4013484) is no longer running 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 1 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # es=1 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 4013291 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@950 -- # '[' -z 4013291 ']' 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # kill -0 4013291 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # uname 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4013291 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4013291' 00:04:28.554 killing process with pid 4013291 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@969 -- # kill 4013291 00:04:28.554 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@974 -- # wait 4013291 00:04:28.816 00:04:28.816 real 0m1.789s 00:04:28.816 user 0m5.206s 00:04:28.816 sys 0m0.382s 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:04:28.816 ************************************ 00:04:28.816 END TEST locking_overlapped_coremask 00:04:28.816 ************************************ 00:04:28.816 17:25:36 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:04:28.816 17:25:36 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:28.816 17:25:36 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:28.816 17:25:36 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:28.816 ************************************ 00:04:28.816 START TEST locking_overlapped_coremask_via_rpc 00:04:28.816 ************************************ 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask_via_rpc 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=4013667 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 4013667 /var/tmp/spdk.sock 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 4013667 ']' 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:28.816 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:28.816 17:25:36 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:28.816 [2024-10-17 17:25:36.639836] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:28.816 [2024-10-17 17:25:36.639920] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4013667 ] 00:04:28.816 [2024-10-17 17:25:36.721839] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:04:28.816 [2024-10-17 17:25:36.721867] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:04:29.076 [2024-10-17 17:25:36.759727] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:29.076 [2024-10-17 17:25:36.759944] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:29.076 [2024-10-17 17:25:36.759945] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=4013930 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 4013930 /var/tmp/spdk2.sock 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 4013930 ']' 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:04:29.648 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:29.648 17:25:37 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:29.648 [2024-10-17 17:25:37.479714] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:29.648 [2024-10-17 17:25:37.479772] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4013930 ] 00:04:29.908 [2024-10-17 17:25:37.575382] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:04:29.908 [2024-10-17 17:25:37.575411] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:04:29.908 [2024-10-17 17:25:37.649920] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:04:29.908 [2024-10-17 17:25:37.650078] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:04:29.908 [2024-10-17 17:25:37.650079] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # local es=0 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:30.480 [2024-10-17 17:25:38.282778] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 4013667 has claimed it. 00:04:30.480 request: 00:04:30.480 { 00:04:30.480 "method": "framework_enable_cpumask_locks", 00:04:30.480 "req_id": 1 00:04:30.480 } 00:04:30.480 Got JSON-RPC error response 00:04:30.480 response: 00:04:30.480 { 00:04:30.480 "code": -32603, 00:04:30.480 "message": "Failed to claim CPU core: 2" 00:04:30.480 } 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # es=1 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 4013667 /var/tmp/spdk.sock 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 4013667 ']' 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:30.480 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:30.480 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 4013930 /var/tmp/spdk2.sock 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 4013930 ']' 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:04:30.741 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:04:30.741 00:04:30.741 real 0m2.080s 00:04:30.741 user 0m0.870s 00:04:30.741 sys 0m0.131s 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:30.741 17:25:38 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:30.741 ************************************ 00:04:30.741 END TEST locking_overlapped_coremask_via_rpc 00:04:30.741 ************************************ 00:04:31.003 17:25:38 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:04:31.003 17:25:38 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 4013667 ]] 00:04:31.003 17:25:38 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 4013667 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 4013667 ']' 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 4013667 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4013667 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4013667' 00:04:31.003 killing process with pid 4013667 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 4013667 00:04:31.003 17:25:38 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 4013667 00:04:31.264 17:25:38 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 4013930 ]] 00:04:31.264 17:25:38 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 4013930 00:04:31.264 17:25:38 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 4013930 ']' 00:04:31.264 17:25:38 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 4013930 00:04:31.264 17:25:38 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:04:31.264 17:25:38 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:31.264 17:25:38 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4013930 00:04:31.264 17:25:39 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:04:31.264 17:25:39 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:04:31.264 17:25:39 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4013930' 00:04:31.264 killing process with pid 4013930 00:04:31.264 17:25:39 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 4013930 00:04:31.264 17:25:39 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 4013930 00:04:31.525 17:25:39 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:04:31.526 17:25:39 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:04:31.526 17:25:39 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 4013667 ]] 00:04:31.526 17:25:39 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 4013667 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 4013667 ']' 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 4013667 00:04:31.526 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (4013667) - No such process 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 4013667 is not found' 00:04:31.526 Process with pid 4013667 is not found 00:04:31.526 17:25:39 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 4013930 ]] 00:04:31.526 17:25:39 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 4013930 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 4013930 ']' 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 4013930 00:04:31.526 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (4013930) - No such process 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 4013930 is not found' 00:04:31.526 Process with pid 4013930 is not found 00:04:31.526 17:25:39 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:04:31.526 00:04:31.526 real 0m15.933s 00:04:31.526 user 0m27.988s 00:04:31.526 sys 0m5.000s 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:31.526 17:25:39 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:04:31.526 ************************************ 00:04:31.526 END TEST cpu_locks 00:04:31.526 ************************************ 00:04:31.526 00:04:31.526 real 0m41.815s 00:04:31.526 user 1m22.272s 00:04:31.526 sys 0m8.371s 00:04:31.526 17:25:39 event -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:31.526 17:25:39 event -- common/autotest_common.sh@10 -- # set +x 00:04:31.526 ************************************ 00:04:31.526 END TEST event 00:04:31.526 ************************************ 00:04:31.526 17:25:39 -- spdk/autotest.sh@169 -- # run_test thread /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:04:31.526 17:25:39 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:31.526 17:25:39 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:31.526 17:25:39 -- common/autotest_common.sh@10 -- # set +x 00:04:31.526 ************************************ 00:04:31.526 START TEST thread 00:04:31.526 ************************************ 00:04:31.526 17:25:39 thread -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:04:31.526 * Looking for test storage... 00:04:31.526 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread 00:04:31.526 17:25:39 thread -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:31.526 17:25:39 thread -- common/autotest_common.sh@1691 -- # lcov --version 00:04:31.526 17:25:39 thread -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:31.787 17:25:39 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:31.787 17:25:39 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:31.787 17:25:39 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:31.787 17:25:39 thread -- scripts/common.sh@336 -- # IFS=.-: 00:04:31.787 17:25:39 thread -- scripts/common.sh@336 -- # read -ra ver1 00:04:31.787 17:25:39 thread -- scripts/common.sh@337 -- # IFS=.-: 00:04:31.787 17:25:39 thread -- scripts/common.sh@337 -- # read -ra ver2 00:04:31.787 17:25:39 thread -- scripts/common.sh@338 -- # local 'op=<' 00:04:31.787 17:25:39 thread -- scripts/common.sh@340 -- # ver1_l=2 00:04:31.787 17:25:39 thread -- scripts/common.sh@341 -- # ver2_l=1 00:04:31.787 17:25:39 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:31.787 17:25:39 thread -- scripts/common.sh@344 -- # case "$op" in 00:04:31.787 17:25:39 thread -- scripts/common.sh@345 -- # : 1 00:04:31.787 17:25:39 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:31.787 17:25:39 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:31.787 17:25:39 thread -- scripts/common.sh@365 -- # decimal 1 00:04:31.787 17:25:39 thread -- scripts/common.sh@353 -- # local d=1 00:04:31.787 17:25:39 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:31.787 17:25:39 thread -- scripts/common.sh@355 -- # echo 1 00:04:31.787 17:25:39 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:04:31.787 17:25:39 thread -- scripts/common.sh@366 -- # decimal 2 00:04:31.787 17:25:39 thread -- scripts/common.sh@353 -- # local d=2 00:04:31.787 17:25:39 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:31.787 17:25:39 thread -- scripts/common.sh@355 -- # echo 2 00:04:31.787 17:25:39 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:04:31.787 17:25:39 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:31.787 17:25:39 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:31.787 17:25:39 thread -- scripts/common.sh@368 -- # return 0 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:31.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:31.787 --rc genhtml_branch_coverage=1 00:04:31.787 --rc genhtml_function_coverage=1 00:04:31.787 --rc genhtml_legend=1 00:04:31.787 --rc geninfo_all_blocks=1 00:04:31.787 --rc geninfo_unexecuted_blocks=1 00:04:31.787 00:04:31.787 ' 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:31.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:31.787 --rc genhtml_branch_coverage=1 00:04:31.787 --rc genhtml_function_coverage=1 00:04:31.787 --rc genhtml_legend=1 00:04:31.787 --rc geninfo_all_blocks=1 00:04:31.787 --rc geninfo_unexecuted_blocks=1 00:04:31.787 00:04:31.787 ' 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:31.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:31.787 --rc genhtml_branch_coverage=1 00:04:31.787 --rc genhtml_function_coverage=1 00:04:31.787 --rc genhtml_legend=1 00:04:31.787 --rc geninfo_all_blocks=1 00:04:31.787 --rc geninfo_unexecuted_blocks=1 00:04:31.787 00:04:31.787 ' 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:31.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:31.787 --rc genhtml_branch_coverage=1 00:04:31.787 --rc genhtml_function_coverage=1 00:04:31.787 --rc genhtml_legend=1 00:04:31.787 --rc geninfo_all_blocks=1 00:04:31.787 --rc geninfo_unexecuted_blocks=1 00:04:31.787 00:04:31.787 ' 00:04:31.787 17:25:39 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:31.787 17:25:39 thread -- common/autotest_common.sh@10 -- # set +x 00:04:31.787 ************************************ 00:04:31.787 START TEST thread_poller_perf 00:04:31.787 ************************************ 00:04:31.787 17:25:39 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:04:31.787 [2024-10-17 17:25:39.585988] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:31.787 [2024-10-17 17:25:39.586073] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4014442 ] 00:04:31.787 [2024-10-17 17:25:39.670212] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:32.048 [2024-10-17 17:25:39.706449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:32.048 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:04:32.988 [2024-10-17T15:25:40.907Z] ====================================== 00:04:32.988 [2024-10-17T15:25:40.907Z] busy:2408161538 (cyc) 00:04:32.988 [2024-10-17T15:25:40.907Z] total_run_count: 417000 00:04:32.988 [2024-10-17T15:25:40.907Z] tsc_hz: 2400000000 (cyc) 00:04:32.988 [2024-10-17T15:25:40.907Z] ====================================== 00:04:32.988 [2024-10-17T15:25:40.907Z] poller_cost: 5774 (cyc), 2405 (nsec) 00:04:32.988 00:04:32.988 real 0m1.177s 00:04:32.988 user 0m1.083s 00:04:32.988 sys 0m0.089s 00:04:32.988 17:25:40 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:32.988 17:25:40 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:04:32.988 ************************************ 00:04:32.988 END TEST thread_poller_perf 00:04:32.988 ************************************ 00:04:32.988 17:25:40 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:04:32.988 17:25:40 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:04:32.988 17:25:40 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:32.988 17:25:40 thread -- common/autotest_common.sh@10 -- # set +x 00:04:32.988 ************************************ 00:04:32.988 START TEST thread_poller_perf 00:04:32.988 ************************************ 00:04:32.988 17:25:40 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:04:32.988 [2024-10-17 17:25:40.840351] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:32.988 [2024-10-17 17:25:40.840465] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4014750 ] 00:04:33.248 [2024-10-17 17:25:40.923539] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:33.248 [2024-10-17 17:25:40.955278] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:33.248 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:04:34.188 [2024-10-17T15:25:42.107Z] ====================================== 00:04:34.188 [2024-10-17T15:25:42.107Z] busy:2401719678 (cyc) 00:04:34.188 [2024-10-17T15:25:42.107Z] total_run_count: 5113000 00:04:34.188 [2024-10-17T15:25:42.107Z] tsc_hz: 2400000000 (cyc) 00:04:34.188 [2024-10-17T15:25:42.107Z] ====================================== 00:04:34.188 [2024-10-17T15:25:42.107Z] poller_cost: 469 (cyc), 195 (nsec) 00:04:34.188 00:04:34.188 real 0m1.164s 00:04:34.188 user 0m1.084s 00:04:34.188 sys 0m0.077s 00:04:34.188 17:25:41 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:34.188 17:25:41 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:04:34.188 ************************************ 00:04:34.188 END TEST thread_poller_perf 00:04:34.188 ************************************ 00:04:34.188 17:25:42 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:04:34.188 00:04:34.188 real 0m2.695s 00:04:34.188 user 0m2.333s 00:04:34.188 sys 0m0.375s 00:04:34.188 17:25:42 thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:34.188 17:25:42 thread -- common/autotest_common.sh@10 -- # set +x 00:04:34.188 ************************************ 00:04:34.188 END TEST thread 00:04:34.188 ************************************ 00:04:34.188 17:25:42 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:04:34.188 17:25:42 -- spdk/autotest.sh@176 -- # run_test app_cmdline /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:04:34.188 17:25:42 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:34.188 17:25:42 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:34.188 17:25:42 -- common/autotest_common.sh@10 -- # set +x 00:04:34.188 ************************************ 00:04:34.188 START TEST app_cmdline 00:04:34.188 ************************************ 00:04:34.188 17:25:42 app_cmdline -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:04:34.449 * Looking for test storage... 00:04:34.449 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:04:34.449 17:25:42 app_cmdline -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:34.449 17:25:42 app_cmdline -- common/autotest_common.sh@1691 -- # lcov --version 00:04:34.449 17:25:42 app_cmdline -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:34.449 17:25:42 app_cmdline -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@345 -- # : 1 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:34.449 17:25:42 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:34.450 17:25:42 app_cmdline -- scripts/common.sh@368 -- # return 0 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:34.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:34.450 --rc genhtml_branch_coverage=1 00:04:34.450 --rc genhtml_function_coverage=1 00:04:34.450 --rc genhtml_legend=1 00:04:34.450 --rc geninfo_all_blocks=1 00:04:34.450 --rc geninfo_unexecuted_blocks=1 00:04:34.450 00:04:34.450 ' 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:34.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:34.450 --rc genhtml_branch_coverage=1 00:04:34.450 --rc genhtml_function_coverage=1 00:04:34.450 --rc genhtml_legend=1 00:04:34.450 --rc geninfo_all_blocks=1 00:04:34.450 --rc geninfo_unexecuted_blocks=1 00:04:34.450 00:04:34.450 ' 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:34.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:34.450 --rc genhtml_branch_coverage=1 00:04:34.450 --rc genhtml_function_coverage=1 00:04:34.450 --rc genhtml_legend=1 00:04:34.450 --rc geninfo_all_blocks=1 00:04:34.450 --rc geninfo_unexecuted_blocks=1 00:04:34.450 00:04:34.450 ' 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:34.450 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:34.450 --rc genhtml_branch_coverage=1 00:04:34.450 --rc genhtml_function_coverage=1 00:04:34.450 --rc genhtml_legend=1 00:04:34.450 --rc geninfo_all_blocks=1 00:04:34.450 --rc geninfo_unexecuted_blocks=1 00:04:34.450 00:04:34.450 ' 00:04:34.450 17:25:42 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:04:34.450 17:25:42 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=4015036 00:04:34.450 17:25:42 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 4015036 00:04:34.450 17:25:42 app_cmdline -- app/cmdline.sh@16 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@831 -- # '[' -z 4015036 ']' 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:34.450 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:34.450 17:25:42 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:04:34.450 [2024-10-17 17:25:42.356048] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:34.450 [2024-10-17 17:25:42.356108] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4015036 ] 00:04:34.710 [2024-10-17 17:25:42.435892] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:34.710 [2024-10-17 17:25:42.474347] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:04:35.280 17:25:43 app_cmdline -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:35.280 17:25:43 app_cmdline -- common/autotest_common.sh@864 -- # return 0 00:04:35.280 17:25:43 app_cmdline -- app/cmdline.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py spdk_get_version 00:04:35.540 { 00:04:35.540 "version": "SPDK v25.01-pre git sha1 18d1cad7b", 00:04:35.540 "fields": { 00:04:35.540 "major": 25, 00:04:35.540 "minor": 1, 00:04:35.540 "patch": 0, 00:04:35.540 "suffix": "-pre", 00:04:35.540 "commit": "18d1cad7b" 00:04:35.540 } 00:04:35.540 } 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@26 -- # sort 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:04:35.540 17:25:43 app_cmdline -- app/cmdline.sh@30 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@650 -- # local es=0 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:35.540 17:25:43 app_cmdline -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:04:35.541 17:25:43 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:35.541 17:25:43 app_cmdline -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:04:35.541 17:25:43 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:04:35.541 17:25:43 app_cmdline -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:04:35.541 17:25:43 app_cmdline -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:04:35.541 17:25:43 app_cmdline -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:04:35.801 request: 00:04:35.801 { 00:04:35.801 "method": "env_dpdk_get_mem_stats", 00:04:35.801 "req_id": 1 00:04:35.801 } 00:04:35.801 Got JSON-RPC error response 00:04:35.801 response: 00:04:35.801 { 00:04:35.801 "code": -32601, 00:04:35.801 "message": "Method not found" 00:04:35.801 } 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@653 -- # es=1 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:04:35.801 17:25:43 app_cmdline -- app/cmdline.sh@1 -- # killprocess 4015036 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@950 -- # '[' -z 4015036 ']' 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@954 -- # kill -0 4015036 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@955 -- # uname 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4015036 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4015036' 00:04:35.801 killing process with pid 4015036 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@969 -- # kill 4015036 00:04:35.801 17:25:43 app_cmdline -- common/autotest_common.sh@974 -- # wait 4015036 00:04:36.061 00:04:36.061 real 0m1.734s 00:04:36.061 user 0m2.084s 00:04:36.061 sys 0m0.468s 00:04:36.061 17:25:43 app_cmdline -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:36.061 17:25:43 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:04:36.061 ************************************ 00:04:36.061 END TEST app_cmdline 00:04:36.061 ************************************ 00:04:36.061 17:25:43 -- spdk/autotest.sh@177 -- # run_test version /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:04:36.061 17:25:43 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:04:36.061 17:25:43 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:36.061 17:25:43 -- common/autotest_common.sh@10 -- # set +x 00:04:36.061 ************************************ 00:04:36.061 START TEST version 00:04:36.061 ************************************ 00:04:36.061 17:25:43 version -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:04:36.322 * Looking for test storage... 00:04:36.322 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:04:36.322 17:25:44 version -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:36.322 17:25:44 version -- common/autotest_common.sh@1691 -- # lcov --version 00:04:36.322 17:25:44 version -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:36.322 17:25:44 version -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:36.322 17:25:44 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:36.322 17:25:44 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:36.322 17:25:44 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:36.323 17:25:44 version -- scripts/common.sh@336 -- # IFS=.-: 00:04:36.323 17:25:44 version -- scripts/common.sh@336 -- # read -ra ver1 00:04:36.323 17:25:44 version -- scripts/common.sh@337 -- # IFS=.-: 00:04:36.323 17:25:44 version -- scripts/common.sh@337 -- # read -ra ver2 00:04:36.323 17:25:44 version -- scripts/common.sh@338 -- # local 'op=<' 00:04:36.323 17:25:44 version -- scripts/common.sh@340 -- # ver1_l=2 00:04:36.323 17:25:44 version -- scripts/common.sh@341 -- # ver2_l=1 00:04:36.323 17:25:44 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:36.323 17:25:44 version -- scripts/common.sh@344 -- # case "$op" in 00:04:36.323 17:25:44 version -- scripts/common.sh@345 -- # : 1 00:04:36.323 17:25:44 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:36.323 17:25:44 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:36.323 17:25:44 version -- scripts/common.sh@365 -- # decimal 1 00:04:36.323 17:25:44 version -- scripts/common.sh@353 -- # local d=1 00:04:36.323 17:25:44 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:36.323 17:25:44 version -- scripts/common.sh@355 -- # echo 1 00:04:36.323 17:25:44 version -- scripts/common.sh@365 -- # ver1[v]=1 00:04:36.323 17:25:44 version -- scripts/common.sh@366 -- # decimal 2 00:04:36.323 17:25:44 version -- scripts/common.sh@353 -- # local d=2 00:04:36.323 17:25:44 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:36.323 17:25:44 version -- scripts/common.sh@355 -- # echo 2 00:04:36.323 17:25:44 version -- scripts/common.sh@366 -- # ver2[v]=2 00:04:36.323 17:25:44 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:36.323 17:25:44 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:36.323 17:25:44 version -- scripts/common.sh@368 -- # return 0 00:04:36.323 17:25:44 version -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:36.323 17:25:44 version -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:36.323 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.323 --rc genhtml_branch_coverage=1 00:04:36.323 --rc genhtml_function_coverage=1 00:04:36.323 --rc genhtml_legend=1 00:04:36.323 --rc geninfo_all_blocks=1 00:04:36.323 --rc geninfo_unexecuted_blocks=1 00:04:36.323 00:04:36.323 ' 00:04:36.323 17:25:44 version -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:36.323 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.323 --rc genhtml_branch_coverage=1 00:04:36.323 --rc genhtml_function_coverage=1 00:04:36.323 --rc genhtml_legend=1 00:04:36.323 --rc geninfo_all_blocks=1 00:04:36.323 --rc geninfo_unexecuted_blocks=1 00:04:36.323 00:04:36.323 ' 00:04:36.323 17:25:44 version -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:36.323 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.323 --rc genhtml_branch_coverage=1 00:04:36.323 --rc genhtml_function_coverage=1 00:04:36.323 --rc genhtml_legend=1 00:04:36.323 --rc geninfo_all_blocks=1 00:04:36.323 --rc geninfo_unexecuted_blocks=1 00:04:36.323 00:04:36.323 ' 00:04:36.323 17:25:44 version -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:36.323 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.323 --rc genhtml_branch_coverage=1 00:04:36.323 --rc genhtml_function_coverage=1 00:04:36.323 --rc genhtml_legend=1 00:04:36.323 --rc geninfo_all_blocks=1 00:04:36.323 --rc geninfo_unexecuted_blocks=1 00:04:36.323 00:04:36.323 ' 00:04:36.323 17:25:44 version -- app/version.sh@17 -- # get_header_version major 00:04:36.323 17:25:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # cut -f2 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # tr -d '"' 00:04:36.323 17:25:44 version -- app/version.sh@17 -- # major=25 00:04:36.323 17:25:44 version -- app/version.sh@18 -- # get_header_version minor 00:04:36.323 17:25:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # cut -f2 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # tr -d '"' 00:04:36.323 17:25:44 version -- app/version.sh@18 -- # minor=1 00:04:36.323 17:25:44 version -- app/version.sh@19 -- # get_header_version patch 00:04:36.323 17:25:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # cut -f2 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # tr -d '"' 00:04:36.323 17:25:44 version -- app/version.sh@19 -- # patch=0 00:04:36.323 17:25:44 version -- app/version.sh@20 -- # get_header_version suffix 00:04:36.323 17:25:44 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # cut -f2 00:04:36.323 17:25:44 version -- app/version.sh@14 -- # tr -d '"' 00:04:36.323 17:25:44 version -- app/version.sh@20 -- # suffix=-pre 00:04:36.323 17:25:44 version -- app/version.sh@22 -- # version=25.1 00:04:36.323 17:25:44 version -- app/version.sh@25 -- # (( patch != 0 )) 00:04:36.323 17:25:44 version -- app/version.sh@28 -- # version=25.1rc0 00:04:36.323 17:25:44 version -- app/version.sh@30 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:04:36.323 17:25:44 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:04:36.323 17:25:44 version -- app/version.sh@30 -- # py_version=25.1rc0 00:04:36.323 17:25:44 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:04:36.323 00:04:36.323 real 0m0.286s 00:04:36.323 user 0m0.168s 00:04:36.323 sys 0m0.166s 00:04:36.323 17:25:44 version -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:36.323 17:25:44 version -- common/autotest_common.sh@10 -- # set +x 00:04:36.323 ************************************ 00:04:36.323 END TEST version 00:04:36.323 ************************************ 00:04:36.323 17:25:44 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:04:36.323 17:25:44 -- spdk/autotest.sh@188 -- # [[ 0 -eq 1 ]] 00:04:36.583 17:25:44 -- spdk/autotest.sh@194 -- # uname -s 00:04:36.583 17:25:44 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:04:36.583 17:25:44 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:04:36.583 17:25:44 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:04:36.583 17:25:44 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:04:36.583 17:25:44 -- spdk/autotest.sh@252 -- # '[' 0 -eq 1 ']' 00:04:36.583 17:25:44 -- spdk/autotest.sh@256 -- # timing_exit lib 00:04:36.583 17:25:44 -- common/autotest_common.sh@730 -- # xtrace_disable 00:04:36.584 17:25:44 -- common/autotest_common.sh@10 -- # set +x 00:04:36.584 17:25:44 -- spdk/autotest.sh@258 -- # '[' 0 -eq 1 ']' 00:04:36.584 17:25:44 -- spdk/autotest.sh@263 -- # '[' 0 -eq 1 ']' 00:04:36.584 17:25:44 -- spdk/autotest.sh@272 -- # '[' 1 -eq 1 ']' 00:04:36.584 17:25:44 -- spdk/autotest.sh@273 -- # export NET_TYPE 00:04:36.584 17:25:44 -- spdk/autotest.sh@276 -- # '[' tcp = rdma ']' 00:04:36.584 17:25:44 -- spdk/autotest.sh@279 -- # '[' tcp = tcp ']' 00:04:36.584 17:25:44 -- spdk/autotest.sh@280 -- # run_test nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:04:36.584 17:25:44 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:04:36.584 17:25:44 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:36.584 17:25:44 -- common/autotest_common.sh@10 -- # set +x 00:04:36.584 ************************************ 00:04:36.584 START TEST nvmf_tcp 00:04:36.584 ************************************ 00:04:36.584 17:25:44 nvmf_tcp -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:04:36.584 * Looking for test storage... 00:04:36.584 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:04:36.584 17:25:44 nvmf_tcp -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:36.584 17:25:44 nvmf_tcp -- common/autotest_common.sh@1691 -- # lcov --version 00:04:36.584 17:25:44 nvmf_tcp -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:36.845 17:25:44 nvmf_tcp -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:36.845 17:25:44 nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:04:36.845 17:25:44 nvmf_tcp -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:36.845 17:25:44 nvmf_tcp -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:36.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.845 --rc genhtml_branch_coverage=1 00:04:36.845 --rc genhtml_function_coverage=1 00:04:36.845 --rc genhtml_legend=1 00:04:36.845 --rc geninfo_all_blocks=1 00:04:36.845 --rc geninfo_unexecuted_blocks=1 00:04:36.845 00:04:36.845 ' 00:04:36.845 17:25:44 nvmf_tcp -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:36.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.846 --rc genhtml_branch_coverage=1 00:04:36.846 --rc genhtml_function_coverage=1 00:04:36.846 --rc genhtml_legend=1 00:04:36.846 --rc geninfo_all_blocks=1 00:04:36.846 --rc geninfo_unexecuted_blocks=1 00:04:36.846 00:04:36.846 ' 00:04:36.846 17:25:44 nvmf_tcp -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:36.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.846 --rc genhtml_branch_coverage=1 00:04:36.846 --rc genhtml_function_coverage=1 00:04:36.846 --rc genhtml_legend=1 00:04:36.846 --rc geninfo_all_blocks=1 00:04:36.846 --rc geninfo_unexecuted_blocks=1 00:04:36.846 00:04:36.846 ' 00:04:36.846 17:25:44 nvmf_tcp -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:36.846 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:36.846 --rc genhtml_branch_coverage=1 00:04:36.846 --rc genhtml_function_coverage=1 00:04:36.846 --rc genhtml_legend=1 00:04:36.846 --rc geninfo_all_blocks=1 00:04:36.846 --rc geninfo_unexecuted_blocks=1 00:04:36.846 00:04:36.846 ' 00:04:36.846 17:25:44 nvmf_tcp -- nvmf/nvmf.sh@10 -- # uname -s 00:04:36.846 17:25:44 nvmf_tcp -- nvmf/nvmf.sh@10 -- # '[' '!' Linux = Linux ']' 00:04:36.846 17:25:44 nvmf_tcp -- nvmf/nvmf.sh@14 -- # run_test nvmf_target_core /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:04:36.846 17:25:44 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:04:36.846 17:25:44 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:36.846 17:25:44 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:36.846 ************************************ 00:04:36.846 START TEST nvmf_target_core 00:04:36.846 ************************************ 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:04:36.846 * Looking for test storage... 00:04:36.846 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1691 -- # lcov --version 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # IFS=.-: 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # read -ra ver1 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # IFS=.-: 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # read -ra ver2 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@338 -- # local 'op=<' 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@340 -- # ver1_l=2 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@341 -- # ver2_l=1 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@344 -- # case "$op" in 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@345 -- # : 1 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # decimal 1 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=1 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 1 00:04:36.846 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # ver1[v]=1 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # decimal 2 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=2 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 2 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # ver2[v]=2 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # return 0 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:37.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.108 --rc genhtml_branch_coverage=1 00:04:37.108 --rc genhtml_function_coverage=1 00:04:37.108 --rc genhtml_legend=1 00:04:37.108 --rc geninfo_all_blocks=1 00:04:37.108 --rc geninfo_unexecuted_blocks=1 00:04:37.108 00:04:37.108 ' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:37.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.108 --rc genhtml_branch_coverage=1 00:04:37.108 --rc genhtml_function_coverage=1 00:04:37.108 --rc genhtml_legend=1 00:04:37.108 --rc geninfo_all_blocks=1 00:04:37.108 --rc geninfo_unexecuted_blocks=1 00:04:37.108 00:04:37.108 ' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:37.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.108 --rc genhtml_branch_coverage=1 00:04:37.108 --rc genhtml_function_coverage=1 00:04:37.108 --rc genhtml_legend=1 00:04:37.108 --rc geninfo_all_blocks=1 00:04:37.108 --rc geninfo_unexecuted_blocks=1 00:04:37.108 00:04:37.108 ' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:37.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.108 --rc genhtml_branch_coverage=1 00:04:37.108 --rc genhtml_function_coverage=1 00:04:37.108 --rc genhtml_legend=1 00:04:37.108 --rc geninfo_all_blocks=1 00:04:37.108 --rc geninfo_unexecuted_blocks=1 00:04:37.108 00:04:37.108 ' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # uname -s 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@15 -- # shopt -s extglob 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- paths/export.sh@5 -- # export PATH 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@51 -- # : 0 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:37.108 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:04:37.108 ************************************ 00:04:37.108 START TEST nvmf_abort 00:04:37.108 ************************************ 00:04:37.108 17:25:44 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:04:37.109 * Looking for test storage... 00:04:37.109 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:04:37.109 17:25:44 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:37.109 17:25:44 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1691 -- # lcov --version 00:04:37.109 17:25:44 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:37.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.370 --rc genhtml_branch_coverage=1 00:04:37.370 --rc genhtml_function_coverage=1 00:04:37.370 --rc genhtml_legend=1 00:04:37.370 --rc geninfo_all_blocks=1 00:04:37.370 --rc geninfo_unexecuted_blocks=1 00:04:37.370 00:04:37.370 ' 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:37.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.370 --rc genhtml_branch_coverage=1 00:04:37.370 --rc genhtml_function_coverage=1 00:04:37.370 --rc genhtml_legend=1 00:04:37.370 --rc geninfo_all_blocks=1 00:04:37.370 --rc geninfo_unexecuted_blocks=1 00:04:37.370 00:04:37.370 ' 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:37.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.370 --rc genhtml_branch_coverage=1 00:04:37.370 --rc genhtml_function_coverage=1 00:04:37.370 --rc genhtml_legend=1 00:04:37.370 --rc geninfo_all_blocks=1 00:04:37.370 --rc geninfo_unexecuted_blocks=1 00:04:37.370 00:04:37.370 ' 00:04:37.370 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:37.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:37.370 --rc genhtml_branch_coverage=1 00:04:37.370 --rc genhtml_function_coverage=1 00:04:37.370 --rc genhtml_legend=1 00:04:37.370 --rc geninfo_all_blocks=1 00:04:37.371 --rc geninfo_unexecuted_blocks=1 00:04:37.371 00:04:37.371 ' 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:37.371 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@474 -- # prepare_net_devs 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@436 -- # local -g is_hw=no 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@438 -- # remove_spdk_ns 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:04:37.371 17:25:45 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:04:45.530 Found 0000:31:00.0 (0x8086 - 0x159b) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:04:45.530 Found 0000:31:00.1 (0x8086 - 0x159b) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:04:45.530 Found net devices under 0000:31:00.0: cvl_0_0 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:04:45.530 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:04:45.531 Found net devices under 0000:31:00.1: cvl_0_1 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # is_hw=yes 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:04:45.531 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:04:45.531 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.637 ms 00:04:45.531 00:04:45.531 --- 10.0.0.2 ping statistics --- 00:04:45.531 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:04:45.531 rtt min/avg/max/mdev = 0.637/0.637/0.637/0.000 ms 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:04:45.531 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:04:45.531 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.316 ms 00:04:45.531 00:04:45.531 --- 10.0.0.1 ping statistics --- 00:04:45.531 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:04:45.531 rtt min/avg/max/mdev = 0.316/0.316/0.316/0.000 ms 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@448 -- # return 0 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@724 -- # xtrace_disable 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@507 -- # nvmfpid=4019500 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@508 -- # waitforlisten 4019500 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@831 -- # '[' -z 4019500 ']' 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:45.531 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:45.531 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.531 [2024-10-17 17:25:52.732314] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:45.531 [2024-10-17 17:25:52.732378] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:04:45.531 [2024-10-17 17:25:52.821215] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:04:45.531 [2024-10-17 17:25:52.876260] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:04:45.531 [2024-10-17 17:25:52.876310] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:04:45.531 [2024-10-17 17:25:52.876319] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:04:45.531 [2024-10-17 17:25:52.876326] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:04:45.531 [2024-10-17 17:25:52.876333] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:04:45.531 [2024-10-17 17:25:52.878272] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:04:45.531 [2024-10-17 17:25:52.878431] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:45.531 [2024-10-17 17:25:52.878432] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@864 -- # return 0 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@730 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 [2024-10-17 17:25:53.611993] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 Malloc0 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 Delay0 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 [2024-10-17 17:25:53.694277] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:45.794 17:25:53 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:04:46.055 [2024-10-17 17:25:53.822892] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:04:48.600 Initializing NVMe Controllers 00:04:48.600 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:04:48.600 controller IO queue size 128 less than required 00:04:48.600 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:04:48.600 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:04:48.600 Initialization complete. Launching workers. 00:04:48.600 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 28432 00:04:48.600 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 28493, failed to submit 62 00:04:48.600 success 28436, unsuccessful 57, failed 0 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@514 -- # nvmfcleanup 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:04:48.600 rmmod nvme_tcp 00:04:48.600 rmmod nvme_fabrics 00:04:48.600 rmmod nvme_keyring 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@515 -- # '[' -n 4019500 ']' 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@516 -- # killprocess 4019500 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@950 -- # '[' -z 4019500 ']' 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@954 -- # kill -0 4019500 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@955 -- # uname 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4019500 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4019500' 00:04:48.600 killing process with pid 4019500 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@969 -- # kill 4019500 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@974 -- # wait 4019500 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@789 -- # iptables-save 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@789 -- # iptables-restore 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:04:48.600 17:25:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:04:50.510 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:04:50.510 00:04:50.510 real 0m13.529s 00:04:50.510 user 0m14.357s 00:04:50.510 sys 0m6.659s 00:04:50.510 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:04:50.510 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:04:50.510 ************************************ 00:04:50.510 END TEST nvmf_abort 00:04:50.510 ************************************ 00:04:50.510 17:25:58 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:04:50.510 17:25:58 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:04:50.510 17:25:58 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:04:50.510 17:25:58 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:04:50.772 ************************************ 00:04:50.772 START TEST nvmf_ns_hotplug_stress 00:04:50.772 ************************************ 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:04:50.772 * Looking for test storage... 00:04:50.772 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1691 -- # lcov --version 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:04:50.772 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.772 --rc genhtml_branch_coverage=1 00:04:50.772 --rc genhtml_function_coverage=1 00:04:50.772 --rc genhtml_legend=1 00:04:50.772 --rc geninfo_all_blocks=1 00:04:50.772 --rc geninfo_unexecuted_blocks=1 00:04:50.772 00:04:50.772 ' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:04:50.772 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.772 --rc genhtml_branch_coverage=1 00:04:50.772 --rc genhtml_function_coverage=1 00:04:50.772 --rc genhtml_legend=1 00:04:50.772 --rc geninfo_all_blocks=1 00:04:50.772 --rc geninfo_unexecuted_blocks=1 00:04:50.772 00:04:50.772 ' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:04:50.772 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.772 --rc genhtml_branch_coverage=1 00:04:50.772 --rc genhtml_function_coverage=1 00:04:50.772 --rc genhtml_legend=1 00:04:50.772 --rc geninfo_all_blocks=1 00:04:50.772 --rc geninfo_unexecuted_blocks=1 00:04:50.772 00:04:50.772 ' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:04:50.772 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:50.772 --rc genhtml_branch_coverage=1 00:04:50.772 --rc genhtml_function_coverage=1 00:04:50.772 --rc genhtml_legend=1 00:04:50.772 --rc geninfo_all_blocks=1 00:04:50.772 --rc geninfo_unexecuted_blocks=1 00:04:50.772 00:04:50.772 ' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:50.772 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:50.773 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:50.773 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:50.773 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:50.773 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:50.773 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:50.773 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:50.773 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # prepare_net_devs 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@436 -- # local -g is_hw=no 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # remove_spdk_ns 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:04:51.033 17:25:58 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:04:59.189 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:04:59.190 Found 0000:31:00.0 (0x8086 - 0x159b) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:04:59.190 Found 0000:31:00.1 (0x8086 - 0x159b) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:04:59.190 Found net devices under 0000:31:00.0: cvl_0_0 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:04:59.190 Found net devices under 0000:31:00.1: cvl_0_1 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # is_hw=yes 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:04:59.190 17:26:05 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:04:59.190 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:04:59.190 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.617 ms 00:04:59.190 00:04:59.190 --- 10.0.0.2 ping statistics --- 00:04:59.190 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:04:59.190 rtt min/avg/max/mdev = 0.617/0.617/0.617/0.000 ms 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:04:59.190 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:04:59.190 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.284 ms 00:04:59.190 00:04:59.190 --- 10.0.0.1 ping statistics --- 00:04:59.190 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:04:59.190 rtt min/avg/max/mdev = 0.284/0.284/0.284/0.000 ms 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@448 -- # return 0 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@724 -- # xtrace_disable 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # nvmfpid=4024474 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # waitforlisten 4024474 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@831 -- # '[' -z 4024474 ']' 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@836 -- # local max_retries=100 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:59.190 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # xtrace_disable 00:04:59.190 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:04:59.190 [2024-10-17 17:26:06.333095] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:04:59.190 [2024-10-17 17:26:06.333165] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:04:59.190 [2024-10-17 17:26:06.428517] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:04:59.190 [2024-10-17 17:26:06.481444] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:04:59.190 [2024-10-17 17:26:06.481495] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:04:59.190 [2024-10-17 17:26:06.481504] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:04:59.190 [2024-10-17 17:26:06.481511] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:04:59.190 [2024-10-17 17:26:06.481517] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:04:59.190 [2024-10-17 17:26:06.483181] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:04:59.190 [2024-10-17 17:26:06.483378] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:04:59.190 [2024-10-17 17:26:06.483379] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # return 0 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@730 -- # xtrace_disable 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:04:59.454 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:04:59.454 [2024-10-17 17:26:07.366229] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:59.716 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:04:59.716 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:04:59.977 [2024-10-17 17:26:07.761430] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:04:59.977 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:05:00.238 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:05:00.499 Malloc0 00:05:00.499 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:05:00.499 Delay0 00:05:00.499 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:00.761 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:05:01.021 NULL1 00:05:01.021 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:05:01.282 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:05:01.282 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=4025174 00:05:01.282 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:01.282 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:01.282 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:01.541 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:05:01.541 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:05:01.801 true 00:05:01.801 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:01.801 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:01.802 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:02.062 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:05:02.062 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:05:02.323 true 00:05:02.323 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:02.323 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:02.323 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:02.584 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:05:02.584 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:05:02.844 true 00:05:02.844 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:02.844 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:03.105 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:03.105 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:05:03.105 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:05:03.365 true 00:05:03.365 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:03.365 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:03.626 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:03.626 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:05:03.626 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:05:03.887 true 00:05:03.887 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:03.887 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:04.148 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:04.148 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:05:04.148 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:05:04.409 true 00:05:04.409 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:04.409 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:04.671 Read completed with error (sct=0, sc=11) 00:05:04.671 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:04.671 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:04.671 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:04.671 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:04.671 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:04.671 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:04.671 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:04.671 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:05:04.671 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:05:04.932 true 00:05:04.932 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:04.932 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:05.875 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:05.875 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:05:05.875 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:05:06.137 true 00:05:06.137 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:06.137 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:06.398 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:06.398 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:05:06.398 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:05:06.659 true 00:05:06.659 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:06.659 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:06.920 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:06.920 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:05:06.920 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:05:07.181 true 00:05:07.181 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:07.181 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:07.442 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:07.442 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:05:07.442 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:05:07.704 true 00:05:07.704 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:07.704 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:09.086 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:09.086 17:26:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:09.087 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:09.087 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:09.087 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:09.087 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:09.087 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:09.087 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:09.087 17:26:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:05:09.087 17:26:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:05:09.348 true 00:05:09.348 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:09.348 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:10.330 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:10.330 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:05:10.330 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:05:10.330 true 00:05:10.590 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:10.590 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:10.590 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:10.852 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:05:10.852 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:05:11.113 true 00:05:11.113 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:11.113 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:12.058 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:12.058 17:26:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:12.058 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:12.058 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:12.058 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:12.320 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:12.320 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:12.320 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:12.320 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:05:12.320 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:05:12.581 true 00:05:12.581 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:12.581 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:13.523 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:13.523 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:05:13.523 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:05:13.523 true 00:05:13.785 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:13.785 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:13.785 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:14.045 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:05:14.045 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:05:14.305 true 00:05:14.305 17:26:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:14.305 17:26:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:15.245 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.245 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.245 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:15.245 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.245 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.505 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.505 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.505 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.505 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:15.505 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:05:15.505 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:05:15.767 true 00:05:15.767 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:15.767 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:16.710 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:16.710 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:05:16.710 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:05:16.710 true 00:05:16.970 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:16.970 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:16.970 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:17.231 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:05:17.231 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:05:17.492 true 00:05:17.492 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:17.492 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:17.492 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:17.492 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:17.492 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:17.492 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:17.781 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:17.781 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:17.781 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:17.781 [2024-10-17 17:26:25.507571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.507977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.508008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.508040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.781 [2024-10-17 17:26:25.508068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.508982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:17.782 [2024-10-17 17:26:25.509647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.509980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.510981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.511996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.782 [2024-10-17 17:26:25.512505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.512973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.513972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.514999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.515994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.516998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.783 [2024-10-17 17:26:25.517027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.517999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.518979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.519976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.520974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.521981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.522013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.522046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.522078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.784 [2024-10-17 17:26:25.522110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.522972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.523991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.524988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.525972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.526004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.526034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.526064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.785 [2024-10-17 17:26:25.526096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.526995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.527948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.528994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529385] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.529976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.530866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.531997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.786 [2024-10-17 17:26:25.532695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.532994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.533991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.534927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.535986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.536999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.537785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.787 [2024-10-17 17:26:25.538816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.538845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.538875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.538913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.538950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.538977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.539977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.540970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.541916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.542505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:05:17.788 [2024-10-17 17:26:25.543155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:05:17.788 [2024-10-17 17:26:25.543497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.543973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.544979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 Message suppressed 999 times: [2024-10-17 17:26:25.545491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 Read completed with error (sct=0, sc=15) 00:05:17.788 [2024-10-17 17:26:25.545519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.788 [2024-10-17 17:26:25.545641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.545977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.546969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.547982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548257] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.548998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.549596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.550999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.789 [2024-10-17 17:26:25.551736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.551988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.552858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553700] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.553989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.554966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.555864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.556990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.557998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.558032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.558062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.558091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.790 [2024-10-17 17:26:25.558124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.558987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.559894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.560990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.561875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.562997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.563997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.564652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.791 [2024-10-17 17:26:25.565485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.565996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566692] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.566992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.567969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.568999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.569991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.570984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.792 [2024-10-17 17:26:25.571276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.571886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.572975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.573990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.574997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.575990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.576825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.577222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.577254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.577287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.793 [2024-10-17 17:26:25.577319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.577979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.578991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.579976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.580979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.581978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.794 [2024-10-17 17:26:25.582456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.582989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:17.795 [2024-10-17 17:26:25.583152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.583990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.584964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.585991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.586974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.587002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.587031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.587061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.587091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.795 [2024-10-17 17:26:25.587121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.587995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.588774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589738] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.589979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.796 [2024-10-17 17:26:25.590983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.591979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.592989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.593974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.797 [2024-10-17 17:26:25.594740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.594782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.594812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.594844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.594874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.595739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.596983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.597990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.798 [2024-10-17 17:26:25.598858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.598888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.598916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.598946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.598975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.599932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.600999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.799 [2024-10-17 17:26:25.601817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.601846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.601881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.601912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.601940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.601969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.602986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.603985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.604981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.605010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.605038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.605070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.605097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.605126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.605156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.800 [2024-10-17 17:26:25.605191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.605968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.606974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.607977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.608989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.609020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.801 [2024-10-17 17:26:25.609050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.609993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.610985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611738] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.611979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.802 [2024-10-17 17:26:25.612350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.612900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.613987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.614971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.615981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.616010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.616044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.803 [2024-10-17 17:26:25.616075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.616983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.617448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.618973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.804 [2024-10-17 17:26:25.619314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.619992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:17.805 [2024-10-17 17:26:25.620225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.620997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.621976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.805 [2024-10-17 17:26:25.622990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.623973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.624626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.625993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.806 [2024-10-17 17:26:25.626785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.626817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.626849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.626879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.626910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.626938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.626977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.627989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.628981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629700] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.629986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.807 [2024-10-17 17:26:25.630761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.630788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.630818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.630849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.630879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.630912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.630944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.630978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.631977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.632993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.633984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.808 [2024-10-17 17:26:25.634490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.634885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.635656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.636970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.637985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638102] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.809 [2024-10-17 17:26:25.638332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.638987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639385] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.639983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.640992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.641978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.810 [2024-10-17 17:26:25.642379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.642795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.643970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.644996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.645979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.811 [2024-10-17 17:26:25.646625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.646655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.646697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.646729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.646760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.646802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.646832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.646865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.647982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.648981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.649981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.650996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.812 [2024-10-17 17:26:25.651261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.651965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.652968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.653981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.813 [2024-10-17 17:26:25.654676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.654715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.654746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.654789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.654819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.654849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.655981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.656969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:17.814 [2024-10-17 17:26:25.657910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.657988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.658990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.814 [2024-10-17 17:26:25.659275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.659955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.660990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.661983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.662983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.663014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.815 [2024-10-17 17:26:25.663218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.663973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.664997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.665994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.816 [2024-10-17 17:26:25.666717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.666983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.667938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.668995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.817 [2024-10-17 17:26:25.669709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.669969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.670970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.671972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:17.818 [2024-10-17 17:26:25.672520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.108 [2024-10-17 17:26:25.672549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.108 [2024-10-17 17:26:25.672588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.108 [2024-10-17 17:26:25.672617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.108 [2024-10-17 17:26:25.672649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.108 [2024-10-17 17:26:25.672683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.672827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.672862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.672893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.672926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.672956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.672986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.673976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.674879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.675986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.109 [2024-10-17 17:26:25.676471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.676995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.677988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.678983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.679974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.110 [2024-10-17 17:26:25.680432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.680997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681738] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.681997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.682978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.683970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.111 [2024-10-17 17:26:25.684251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.684982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.685977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.686795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.687988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.112 [2024-10-17 17:26:25.688362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.688989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.689969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.690978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.113 [2024-10-17 17:26:25.691340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.691768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.692985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.693990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.114 [2024-10-17 17:26:25.694889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.694988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.114 [2024-10-17 17:26:25.695717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.695978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696257] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.696993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.697977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.698984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.699014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.699046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.699088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.115 [2024-10-17 17:26:25.699447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.699970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 true 00:05:18.116 [2024-10-17 17:26:25.700875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.700975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.701978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.702983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.703014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.703045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.703083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.703111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.703146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.116 [2024-10-17 17:26:25.703176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.703916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.704986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.705984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.706018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.117 [2024-10-17 17:26:25.706049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.706994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.707970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.118 [2024-10-17 17:26:25.708559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.708976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.709985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.710970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.711003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.711033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.711068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.119 [2024-10-17 17:26:25.711097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.711994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.712977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.120 [2024-10-17 17:26:25.713821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.713853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.713887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.713917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.713950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.713986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.714988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.715963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.121 [2024-10-17 17:26:25.716219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.716761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.717985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.122 [2024-10-17 17:26:25.718851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.718883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.718919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.718950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.718980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.719991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.720968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.123 [2024-10-17 17:26:25.721298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.721982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.722999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.723799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.124 [2024-10-17 17:26:25.724823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.724854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.724889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.724926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.724957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.724989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.725985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.726986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.727994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.125 [2024-10-17 17:26:25.728672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.728711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.728743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.728773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.728912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.728942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.728986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:18.126 [2024-10-17 17:26:25.729175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:18.126 [2024-10-17 17:26:25.729740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.729971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.730990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.731996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.126 [2024-10-17 17:26:25.732164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.126 [2024-10-17 17:26:25.732289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.732968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.733993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.734993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.735857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.736011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.736041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.736070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.736112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.736143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.736174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.127 [2024-10-17 17:26:25.736206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.736970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.737982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.738860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.128 [2024-10-17 17:26:25.739641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.739966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.740977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.741830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.742981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.129 [2024-10-17 17:26:25.743546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.743972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.744980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745700] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.745899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.746979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.130 [2024-10-17 17:26:25.747508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.747970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.748992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.749678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.750960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.131 [2024-10-17 17:26:25.751302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.751992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.752995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.753985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.132 [2024-10-17 17:26:25.754824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.754856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.754886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.754920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.754951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.754979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.755993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.756850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.757981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.133 [2024-10-17 17:26:25.758741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.758998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.759974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.760967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.761994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.134 [2024-10-17 17:26:25.762424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.762978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763102] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.763925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.764991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.765982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.135 [2024-10-17 17:26:25.766369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.766995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.767994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.768981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.136 [2024-10-17 17:26:25.769715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.769976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.137 [2024-10-17 17:26:25.770584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.770973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.771997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.772970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.137 [2024-10-17 17:26:25.773507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.773990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.774988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.775665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.776994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.138 [2024-10-17 17:26:25.777264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.777999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.778997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.779986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.139 [2024-10-17 17:26:25.780872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.780903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.780934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.780965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.781979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.782979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.783973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.140 [2024-10-17 17:26:25.784767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.784805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.784838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.784867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.784901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.784934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.784967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.785980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.786984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787385] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.141 [2024-10-17 17:26:25.787629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.787985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.788996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.789988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.790994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.142 [2024-10-17 17:26:25.791655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.791999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.792955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.793979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.794993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.795024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.795057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.795085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.795114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.795150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.143 [2024-10-17 17:26:25.795181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.795975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.796970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.797920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.798976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.799009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.799038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.799069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.799109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.799139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.799170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.144 [2024-10-17 17:26:25.799205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.799984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.800983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.801964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.145 [2024-10-17 17:26:25.802221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.802986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.803986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.804973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.805974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.146 [2024-10-17 17:26:25.806005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.806994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.147 [2024-10-17 17:26:25.807753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.807978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.808978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.147 [2024-10-17 17:26:25.809872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.809904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.809952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.809984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.810993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.811975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.812865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.148 [2024-10-17 17:26:25.813872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.813905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.813937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.813968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.813998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.814987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.815992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.816977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.149 [2024-10-17 17:26:25.817498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.817986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.818982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.819966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.820636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.821185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.821213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.150 [2024-10-17 17:26:25.821245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.821995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.822997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.823993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.151 [2024-10-17 17:26:25.824788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.824814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.824851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.824880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.824910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.824941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.824975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.825974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.826989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.827991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.828640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.829141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.829171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.829210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.829240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.829269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.829298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.152 [2024-10-17 17:26:25.829326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.829993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.830977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.153 [2024-10-17 17:26:25.831890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.831920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.831951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.831983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.832998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.833987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.834986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.154 [2024-10-17 17:26:25.835310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.835903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.836982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.837988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.838955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.839085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.839114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.155 [2024-10-17 17:26:25.839154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.839562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.840980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.841985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.842971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.156 [2024-10-17 17:26:25.843254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.843940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.844978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.845987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.157 [2024-10-17 17:26:25.846459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.846936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.847401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.847432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.847465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.157 [2024-10-17 17:26:25.847497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.847980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.848990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.849983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.158 [2024-10-17 17:26:25.850560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850692] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.850980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.851985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.852991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.853975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.159 [2024-10-17 17:26:25.854713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.854999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.855986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.856995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.857986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.160 [2024-10-17 17:26:25.858438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.858925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.859994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.860997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.861979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.161 [2024-10-17 17:26:25.862458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.862982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.863716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.864999] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.865977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.162 [2024-10-17 17:26:25.866340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.866973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.867983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.868988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869257] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.163 [2024-10-17 17:26:25.869996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870257] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.870997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.871645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.872975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.164 [2024-10-17 17:26:25.873575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.873607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.873828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.873861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.873892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.873922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.873958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.873987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.874993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.875986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.876500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.165 [2024-10-17 17:26:25.877589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.877989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.878984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.879985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.880661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.166 [2024-10-17 17:26:25.881049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.881988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.882983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 [2024-10-17 17:26:25.883425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.167 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.167 [2024-10-17 17:26:25.883456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 [2024-10-17 17:26:25.883739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.168 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:18.168 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:18.168 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:18.168 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:18.168 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:18.454 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:18.454 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:18.454 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:18.454 [2024-10-17 17:26:26.076527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076677] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.454 [2024-10-17 17:26:26.076850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.076879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.076910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.076942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.076971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.077977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.078996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.079976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.455 [2024-10-17 17:26:26.080384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.080971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.081997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082385] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.082957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.456 [2024-10-17 17:26:26.083471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.083505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.083841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.083872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.083899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.083935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.083964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.083994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.084992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.085889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.086024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.086056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.086086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.086117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.457 [2024-10-17 17:26:26.086156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.086527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.087994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.088987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.089932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.458 [2024-10-17 17:26:26.090457] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.090978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.091976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.459 [2024-10-17 17:26:26.092894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.092928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.092982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.093979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.094990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.460 [2024-10-17 17:26:26.095925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.095958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.095983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.096981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.097969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.098985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.099017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.099054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.461 [2024-10-17 17:26:26.099084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.099983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.100974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.101974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.462 [2024-10-17 17:26:26.102362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.102969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.103995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.104982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.105973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.463 [2024-10-17 17:26:26.106004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106690] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.106942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.107989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 17:26:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:05:18.464 [2024-10-17 17:26:26.108313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 17:26:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:05:18.464 [2024-10-17 17:26:26.108683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.108970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.464 [2024-10-17 17:26:26.109401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.109979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.110984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.111993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112027] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.112991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.465 [2024-10-17 17:26:26.113250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113700] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.113982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.114649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.466 [2024-10-17 17:26:26.115028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.115997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.466 [2024-10-17 17:26:26.116962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.116992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.117987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.118988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.119492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.120985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.467 [2024-10-17 17:26:26.121230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.121967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.122707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.123971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.468 [2024-10-17 17:26:26.124968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.125993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.126974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127805] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127933] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.127992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.469 [2024-10-17 17:26:26.128574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.128986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.129986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.130963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.131990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132297] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.470 [2024-10-17 17:26:26.132432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.132921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.133988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.134981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.135958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.471 [2024-10-17 17:26:26.136232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.136954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.137619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.138988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139123] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.139970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.140005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.140038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.140068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.140100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.140135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.140165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.472 [2024-10-17 17:26:26.140314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.140977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141354] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.141771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.142976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.473 [2024-10-17 17:26:26.143647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.143991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.144883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.145993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.146993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147738] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.474 [2024-10-17 17:26:26.147798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.147829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.147859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.147888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.147923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.147950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.147984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.148967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149259] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.149671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.150979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.475 [2024-10-17 17:26:26.151753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.151792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.151821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.151853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.151884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.151919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.151952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.151982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.476 [2024-10-17 17:26:26.152639] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.152959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.153990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.154982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.476 [2024-10-17 17:26:26.155656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.155984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.156988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.157982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.158973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.159004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.477 [2024-10-17 17:26:26.159041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.159974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.160982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.161978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.478 [2024-10-17 17:26:26.162927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.162959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.163968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164891] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.164989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.165971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.479 [2024-10-17 17:26:26.166786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.166819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.166849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.166889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.166916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.166948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.166977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.167974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.168943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.169980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.480 [2024-10-17 17:26:26.170841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.170877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.170908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.170934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.170968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.171993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.172774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173700] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.173950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.481 [2024-10-17 17:26:26.174501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.174992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175751] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.175968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.176982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177290] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177921] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.177981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.178011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.178045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.178077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.482 [2024-10-17 17:26:26.178111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178614] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.178978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.179996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.180927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181762] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.181991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.182019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.182055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.483 [2024-10-17 17:26:26.182085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.182985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183102] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183491] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.183983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184333] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.184993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.484 [2024-10-17 17:26:26.185604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.185640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.185671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.185704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.185735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.185774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.185803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.186978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187559] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.187988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188186] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188864] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.188895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.485 [2024-10-17 17:26:26.189943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.189971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.190994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191057] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191595] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.486 [2024-10-17 17:26:26.191626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191886] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.191986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192768] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192866] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.192971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.193003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.193037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.486 [2024-10-17 17:26:26.193779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.193813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.193843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.193874] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.193907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.193937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.193968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194844] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.194971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195102] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.195992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.196972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197302] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197645] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.487 [2024-10-17 17:26:26.197741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.197994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.198863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199279] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199502] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199849] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.199972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200201] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.200969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201232] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.488 [2024-10-17 17:26:26.201858] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.201884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.201916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.201947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.201979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202575] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.202878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203171] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203445] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203484] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.203546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204872] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.204974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.205962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.206020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.206053] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.489 [2024-10-17 17:26:26.206084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206250] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206413] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206536] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.206992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.207982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208246] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.208971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209809] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.209976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.210012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.210041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.490 [2024-10-17 17:26:26.210078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210235] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.210668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211528] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211827] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.211951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212002] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.212983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.213968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214394] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214560] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214961] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.214993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.491 [2024-10-17 17:26:26.215025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.215788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216260] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.216966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.217983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.218945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.219988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.492 [2024-10-17 17:26:26.220019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220313] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220592] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220701] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220774] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.220995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221411] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221625] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.221992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222090] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222218] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222380] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222447] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222679] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.222979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.223654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.224032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.224068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.493 [2024-10-17 17:26:26.224098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224365] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224772] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224812] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.224974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225267] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225385] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.225975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226102] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.226993] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227091] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227125] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227188] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227842] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227935] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.227970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228187] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228282] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228350] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.494 [2024-10-17 17:26:26.228483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.228512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.228549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.228580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.495 [2024-10-17 17:26:26.228945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.228982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.229978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230245] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.230988] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231442] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231475] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231574] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.231989] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232696] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.232970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233177] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.495 [2024-10-17 17:26:26.233273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.233304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.233338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.233369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.233403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.233437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234214] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234281] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234347] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234820] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.234985] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235113] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235668] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235743] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235908] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.235972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.496 [2024-10-17 17:26:26.236928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.236960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.236998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237285] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237315] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237875] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.237978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238080] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238111] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.238996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239062] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239385] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239714] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.239981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240078] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240112] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240209] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240303] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.497 [2024-10-17 17:26:26.240532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.240987] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241019] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241494] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241765] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241926] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.241991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242086] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242786] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.242994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243226] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243395] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.243997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.244028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.244059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.244117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.244151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.244183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.244219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.498 [2024-10-17 17:26:26.244251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244429] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244464] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244568] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244928] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.244968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245000] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245102] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245162] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245320] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245561] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245737] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.245769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246412] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.246982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247383] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247414] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247478] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247512] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247644] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247776] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247806] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.247971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248034] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248136] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248236] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.499 [2024-10-17 17:26:26.248299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248363] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.248995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249154] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249216] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249308] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249544] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249606] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249675] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249744] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.249980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250052] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250538] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250607] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.250994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251030] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251101] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251229] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251329] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251740] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251802] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251899] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.251997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.252029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.252060] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.252088] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.252121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.252155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.500 [2024-10-17 17:26:26.252193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252422] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252859] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252924] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252958] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.252991] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253455] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253529] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253832] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253956] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.253994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254505] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254930] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254966] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.254998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255059] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255157] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255219] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255346] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.255981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256081] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.501 [2024-10-17 17:26:26.256241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256377] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256473] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256507] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256537] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256585] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256948] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.256995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257061] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257092] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257124] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257158] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257192] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257288] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257384] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.257971] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.258996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259118] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259345] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259376] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259548] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.502 [2024-10-17 17:26:26.259581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259833] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259863] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259893] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259964] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.259995] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260026] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260121] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260152] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260334] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.260975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261038] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261132] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261291] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261520] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261579] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261681] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261783] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261909] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.261972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262068] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262552] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262709] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262876] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.262975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.263781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.263821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.263852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.263882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.263914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.263949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.263980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.264013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.264044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.264077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.264108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.264140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.264172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.503 [2024-10-17 17:26:26.264200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264531] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264686] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.264984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265278] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265440] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265699] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.265828] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266094] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266159] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266191] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266356] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266526] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266724] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.266981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267266] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267423] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 true 00:05:18.504 [2024-10-17 17:26:26.267584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267623] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.504 [2024-10-17 17:26:26.267755] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.267787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.267818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.267855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.267885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.267917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.267957] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.267986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268050] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.505 [2024-10-17 17:26:26.268488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.268975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269287] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269389] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269647] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269777] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269843] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269905] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.269967] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270003] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270263] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.270490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.505 [2024-10-17 17:26:26.271943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.271978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272044] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272074] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272557] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272589] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272685] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272721] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272884] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.272984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273149] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273514] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273578] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273643] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273680] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273939] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.273998] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274025] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274055] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274342] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274479] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274636] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274706] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274808] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.274972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.506 [2024-10-17 17:26:26.275001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275033] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275065] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275128] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275169] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275234] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275953] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.275983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276018] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276231] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276262] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276298] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276401] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276564] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276631] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276697] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276730] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276798] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276937] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.276969] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277001] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277032] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277097] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277257] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277289] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277355] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277386] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277450] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277549] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277713] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.277745] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278310] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278340] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278372] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278468] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278695] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278787] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278911] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.278970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.279009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.279040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.279073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.507 [2024-10-17 17:26:26.279110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279141] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279273] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279305] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279463] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279493] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279591] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279657] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279726] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279759] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279855] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279918] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.279981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280043] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280096] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280129] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280160] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280225] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280318] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280480] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280515] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280638] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280838] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280938] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.280970] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281382] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281416] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281448] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281547] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281648] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281682] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281845] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281878] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281910] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.281975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282105] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282196] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282577] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282674] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282831] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.508 [2024-10-17 17:26:26.282901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.282931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.282962] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.282992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283087] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283156] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283190] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283222] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283255] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283602] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283634] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283711] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283773] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283865] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283896] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283932] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.283992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284028] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284058] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284122] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284153] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284184] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284326] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284361] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284460] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284553] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284583] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284718] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284914] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.284982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285011] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285049] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285304] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.285563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286314] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286344] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286405] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286435] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286499] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286597] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286627] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286658] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286788] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286817] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286885] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.509 [2024-10-17 17:26:26.286941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.286976] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287070] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287106] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287139] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287202] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287322] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287353] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287396] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287426] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287593] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287784] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287915] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287945] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.287977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288010] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288039] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288375] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288408] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288441] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288471] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288563] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288615] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288821] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.288882] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289185] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289283] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289390] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289421] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289483] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289543] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289573] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289611] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289641] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289707] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289738] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289835] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289900] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.289992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290023] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290120] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290151] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290181] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.510 [2024-10-17 17:26:26.290370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290626] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290852] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290943] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.290975] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291063] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291127] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291167] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291228] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291385] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291417] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291449] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291485] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291516] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291613] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291649] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291678] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291712] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291846] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.291983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292016] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292045] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292076] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292168] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292294] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292325] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292420] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292617] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292646] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292676] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292705] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292766] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292797] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.292830] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293220] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293253] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293319] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293381] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293410] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293443] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293476] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293541] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293794] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 17:26:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:18.511 [2024-10-17 17:26:26.293960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.293994] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.294024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.294054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.511 [2024-10-17 17:26:26.294084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294148] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294183] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294286] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 17:26:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:18.512 [2024-10-17 17:26:26.294316] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294357] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294452] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294487] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294517] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294550] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294581] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294612] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294642] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294671] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294739] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294868] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294898] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.294963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295037] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295107] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295137] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295166] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295199] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295230] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295465] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295565] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295599] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295632] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295670] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295738] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295836] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295897] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.295929] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296598] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296629] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296693] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296727] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.296990] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297021] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297175] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297203] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297238] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297337] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297371] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297438] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297503] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297570] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297663] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297702] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297734] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297767] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297801] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.512 [2024-10-17 17:26:26.297867] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.297902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.297931] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.297963] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.297992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298024] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298056] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298085] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298207] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298240] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298656] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298723] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298754] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298785] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298816] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298851] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298950] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.298983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299067] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299093] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299119] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299144] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299170] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299221] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299248] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299352] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299378] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299404] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299430] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299481] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299513] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299546] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299608] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299716] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299778] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299810] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299840] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299873] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299942] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.299973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300103] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300133] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300227] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300261] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300292] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300321] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300351] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300490] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300912] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300944] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.300981] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301042] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301083] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301204] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301243] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301272] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301397] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301427] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.513 [2024-10-17 17:26:26.301461] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301495] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301521] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301757] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301860] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301890] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301922] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301955] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.301986] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302017] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302048] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302079] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302110] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302205] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302237] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302271] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302339] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302369] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302402] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302433] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302533] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302605] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302735] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.302925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303506] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303542] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303571] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303710] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303741] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303803] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303903] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303965] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.303996] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304031] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304064] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304095] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304126] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304155] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304182] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304217] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304252] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304317] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304349] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304415] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304446] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304474] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304508] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304572] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304637] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304672] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304856] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304960] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.304992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305020] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305089] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305150] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305247] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.514 [2024-10-17 17:26:26.305277] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305673] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305708] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305738] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305770] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305804] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305837] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305869] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305902] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.305978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 Message suppressed 999 times: Read completed with error (sct=0, sc=15) 00:05:18.515 [2024-10-17 17:26:26.306009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306114] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306530] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306594] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306717] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306747] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306811] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306847] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306877] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306907] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.306972] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307099] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307200] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307296] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307370] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307467] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307497] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307601] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307666] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307732] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307760] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307789] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307822] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.307954] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308138] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308210] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308336] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308368] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308462] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308650] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.515 [2024-10-17 17:26:26.308719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308813] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308848] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308880] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308946] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.308977] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309007] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309142] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309194] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309254] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309324] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309388] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309419] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309451] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309482] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309511] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309540] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309604] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309664] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309728] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309758] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309792] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309920] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.309978] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310008] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310073] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310135] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310172] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310300] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310327] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310360] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310393] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310458] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310492] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310522] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310558] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310587] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310621] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310653] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310753] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310781] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.310814] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311035] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311069] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311100] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311189] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311215] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311249] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311284] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311312] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311348] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311379] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311409] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311444] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311477] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311510] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311576] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311609] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311659] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311725] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311795] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311825] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.311979] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312015] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312116] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312176] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312274] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312309] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.516 [2024-10-17 17:26:26.312341] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312373] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312403] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312434] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312500] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312906] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312940] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.312974] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313005] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313066] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313098] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313131] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313164] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313197] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313264] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313335] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313437] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313498] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313569] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313665] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313703] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313736] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313769] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313800] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313839] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313870] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313901] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313934] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.313973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314006] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314040] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314109] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314140] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314179] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314208] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314276] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314307] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314338] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314367] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314399] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314436] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314469] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314534] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314624] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314689] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314750] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314782] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314818] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314879] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314916] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.314949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315161] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315193] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315224] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315258] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315331] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315364] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315431] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315466] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315527] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315588] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315619] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.315687] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316195] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316233] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316265] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316358] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316387] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316418] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316453] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316524] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.517 [2024-10-17 17:26:26.316556] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316590] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316622] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316655] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316688] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316722] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316756] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316790] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316887] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316949] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.316982] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317014] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317046] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317075] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317143] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317173] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317211] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317241] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317275] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317306] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317366] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317400] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317432] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317472] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317504] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317535] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317566] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317596] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317628] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317660] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317694] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317729] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317761] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317826] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317861] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317894] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317923] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317952] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.317984] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318013] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318051] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318082] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318115] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318146] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318178] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318212] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318244] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318406] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318509] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318539] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318567] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318633] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318662] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318733] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318764] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318796] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318829] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318892] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.318927] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319009] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319047] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319077] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319108] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319145] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319174] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319206] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319242] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319270] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319299] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319328] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319359] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319398] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319428] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319459] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319496] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319525] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319555] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319586] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319618] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319654] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319720] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319752] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319791] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319823] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319854] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.518 [2024-10-17 17:26:26.319888] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320268] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320301] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320332] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320391] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320425] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320454] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320486] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320523] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320554] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320584] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320620] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320651] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320683] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320749] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320779] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320815] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320850] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320881] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320913] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320941] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.320973] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321004] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321036] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321072] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321104] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321130] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321163] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321198] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321239] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321269] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321295] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321330] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321362] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321392] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321424] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321456] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321489] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321519] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321582] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321616] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321652] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321684] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321719] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321748] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321780] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321819] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321853] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321883] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321917] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321947] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.321980] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322012] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322041] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322071] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322102] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322134] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322165] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322223] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322256] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322293] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322323] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322488] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322551] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322603] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322635] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322667] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322715] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322746] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322775] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322807] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322841] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322871] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322904] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322936] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322968] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.322997] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.323029] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.323518] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.323545] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.323580] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.323610] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.519 [2024-10-17 17:26:26.323640] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323669] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323704] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323742] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323771] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323799] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323834] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323862] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323895] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323925] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323959] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.323992] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324022] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324054] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324084] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324117] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324147] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324180] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324213] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324251] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324280] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324311] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324343] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324374] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324407] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324439] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324470] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324501] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324532] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324562] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324600] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324630] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324661] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324698] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324731] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324763] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324793] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324824] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324857] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324889] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324919] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324951] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:18.520 [2024-10-17 17:26:26.324983] ctrlr_bdev.c: 361:nvmf_bdev_ctrlr_read_cmd: *ERROR*: Read NLB 1 * block size 512 > SGL length 1 00:05:19.464 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:19.464 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:19.464 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:19.726 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:19.726 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:19.726 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:19.726 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:19.726 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:19.726 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:05:19.726 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:05:20.030 true 00:05:20.030 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:20.030 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:20.705 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:20.705 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:20.705 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:20.966 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:05:20.966 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:05:21.227 true 00:05:21.227 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:21.227 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:21.227 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:21.488 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:05:21.488 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:05:21.749 true 00:05:21.749 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:21.749 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:23.135 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:23.135 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:23.135 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:23.135 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:23.135 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:23.135 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:23.135 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:23.135 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:23.135 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:05:23.135 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:05:23.135 true 00:05:23.135 17:26:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:23.135 17:26:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:24.077 17:26:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:24.338 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:05:24.338 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:05:24.338 true 00:05:24.338 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:24.338 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:24.599 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:24.860 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:05:24.860 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:05:24.860 true 00:05:24.860 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:24.860 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:26.244 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:26.244 17:26:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:26.244 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:26.244 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:26.244 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:26.244 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:26.244 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:26.244 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:26.244 17:26:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:05:26.244 17:26:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:05:26.505 true 00:05:26.505 17:26:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:26.506 17:26:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:27.447 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:27.447 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:27.447 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:27.447 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1030 00:05:27.447 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1030 00:05:27.708 true 00:05:27.708 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:27.708 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:27.969 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:27.969 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1031 00:05:27.969 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1031 00:05:28.229 true 00:05:28.229 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:28.229 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:29.616 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:29.616 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:29.616 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:29.616 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:29.616 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:29.616 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:29.616 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:29.616 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:05:29.616 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1032 00:05:29.616 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1032 00:05:29.616 true 00:05:29.616 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:29.616 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:30.559 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:30.827 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1033 00:05:30.827 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1033 00:05:30.827 true 00:05:30.827 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:30.827 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:31.096 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:31.358 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1034 00:05:31.358 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1034 00:05:31.358 true 00:05:31.358 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:31.358 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:32.742 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:32.742 Initializing NVMe Controllers 00:05:32.742 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:05:32.742 Controller IO queue size 128, less than required. 00:05:32.742 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:05:32.742 Controller IO queue size 128, less than required. 00:05:32.742 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:05:32.742 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:05:32.742 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:05:32.742 Initialization complete. Launching workers. 00:05:32.742 ======================================================== 00:05:32.742 Latency(us) 00:05:32.742 Device Information : IOPS MiB/s Average min max 00:05:32.742 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 3075.77 1.50 26450.34 1405.93 1044306.98 00:05:32.742 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 18125.23 8.85 7035.74 1185.02 403182.64 00:05:32.742 ======================================================== 00:05:32.742 Total : 21201.00 10.35 9852.34 1185.02 1044306.98 00:05:32.742 00:05:32.742 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1035 00:05:32.742 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1035 00:05:33.003 true 00:05:33.003 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 4025174 00:05:33.003 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (4025174) - No such process 00:05:33.003 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 4025174 00:05:33.003 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:33.003 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:33.266 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:05:33.266 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:05:33.266 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:05:33.266 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:33.266 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:05:33.527 null0 00:05:33.527 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:33.527 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:33.527 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:05:33.527 null1 00:05:33.527 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:33.527 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:33.527 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:05:33.788 null2 00:05:33.788 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:33.788 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:33.788 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:05:34.049 null3 00:05:34.049 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:34.049 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:34.049 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:05:34.049 null4 00:05:34.049 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:34.049 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:34.049 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:05:34.310 null5 00:05:34.310 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:34.310 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:34.310 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:05:34.570 null6 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:05:34.570 null7 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.570 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 4031885 4031887 4031890 4031893 4031896 4031899 4031902 4031905 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:34.832 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.094 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.356 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:35.618 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:35.880 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:35.881 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:35.881 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:35.881 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:35.881 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:36.141 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:36.141 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.141 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.141 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:36.402 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.403 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.403 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.666 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:36.667 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.927 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:36.928 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.189 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.450 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:37.711 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:37.973 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:05:38.234 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:05:38.234 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:05:38.234 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.234 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.235 17:26:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.235 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@514 -- # nvmfcleanup 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:05:38.495 rmmod nvme_tcp 00:05:38.495 rmmod nvme_fabrics 00:05:38.495 rmmod nvme_keyring 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@515 -- # '[' -n 4024474 ']' 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # killprocess 4024474 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@950 -- # '[' -z 4024474 ']' 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # kill -0 4024474 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # uname 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4024474 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4024474' 00:05:38.495 killing process with pid 4024474 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@969 -- # kill 4024474 00:05:38.495 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@974 -- # wait 4024474 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-save 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-restore 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:05:38.756 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:05:40.669 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:05:40.669 00:05:40.669 real 0m50.075s 00:05:40.669 user 3m14.814s 00:05:40.669 sys 0m15.845s 00:05:40.669 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:40.669 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:05:40.669 ************************************ 00:05:40.669 END TEST nvmf_ns_hotplug_stress 00:05:40.669 ************************************ 00:05:40.669 17:26:48 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:05:40.669 17:26:48 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:05:40.669 17:26:48 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:40.669 17:26:48 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:05:40.931 ************************************ 00:05:40.931 START TEST nvmf_delete_subsystem 00:05:40.931 ************************************ 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:05:40.931 * Looking for test storage... 00:05:40.931 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1691 -- # lcov --version 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:40.931 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:05:40.932 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:40.932 --rc genhtml_branch_coverage=1 00:05:40.932 --rc genhtml_function_coverage=1 00:05:40.932 --rc genhtml_legend=1 00:05:40.932 --rc geninfo_all_blocks=1 00:05:40.932 --rc geninfo_unexecuted_blocks=1 00:05:40.932 00:05:40.932 ' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:05:40.932 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:40.932 --rc genhtml_branch_coverage=1 00:05:40.932 --rc genhtml_function_coverage=1 00:05:40.932 --rc genhtml_legend=1 00:05:40.932 --rc geninfo_all_blocks=1 00:05:40.932 --rc geninfo_unexecuted_blocks=1 00:05:40.932 00:05:40.932 ' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:05:40.932 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:40.932 --rc genhtml_branch_coverage=1 00:05:40.932 --rc genhtml_function_coverage=1 00:05:40.932 --rc genhtml_legend=1 00:05:40.932 --rc geninfo_all_blocks=1 00:05:40.932 --rc geninfo_unexecuted_blocks=1 00:05:40.932 00:05:40.932 ' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:05:40.932 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:40.932 --rc genhtml_branch_coverage=1 00:05:40.932 --rc genhtml_function_coverage=1 00:05:40.932 --rc genhtml_legend=1 00:05:40.932 --rc geninfo_all_blocks=1 00:05:40.932 --rc geninfo_unexecuted_blocks=1 00:05:40.932 00:05:40.932 ' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:40.932 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # prepare_net_devs 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@436 -- # local -g is_hw=no 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # remove_spdk_ns 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:05:40.932 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:05:41.194 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:05:41.194 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:05:41.194 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:05:41.194 17:26:48 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:05:49.343 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:05:49.344 Found 0000:31:00.0 (0x8086 - 0x159b) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:05:49.344 Found 0000:31:00.1 (0x8086 - 0x159b) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:05:49.344 Found net devices under 0000:31:00.0: cvl_0_0 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:05:49.344 Found net devices under 0000:31:00.1: cvl_0_1 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # is_hw=yes 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:05:49.344 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:05:49.344 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.651 ms 00:05:49.344 00:05:49.344 --- 10.0.0.2 ping statistics --- 00:05:49.344 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:05:49.344 rtt min/avg/max/mdev = 0.651/0.651/0.651/0.000 ms 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:05:49.344 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:05:49.344 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.317 ms 00:05:49.344 00:05:49.344 --- 10.0.0.1 ping statistics --- 00:05:49.344 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:05:49.344 rtt min/avg/max/mdev = 0.317/0.317/0.317/0.000 ms 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@448 -- # return 0 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # nvmfpid=4037213 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # waitforlisten 4037213 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@831 -- # '[' -z 4037213 ']' 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:49.344 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:49.344 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.344 [2024-10-17 17:26:56.509517] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:05:49.344 [2024-10-17 17:26:56.509585] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:05:49.344 [2024-10-17 17:26:56.598752] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:49.344 [2024-10-17 17:26:56.650857] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:05:49.344 [2024-10-17 17:26:56.650905] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:05:49.344 [2024-10-17 17:26:56.650914] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:49.345 [2024-10-17 17:26:56.650921] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:49.345 [2024-10-17 17:26:56.650927] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:05:49.345 [2024-10-17 17:26:56.652741] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:05:49.345 [2024-10-17 17:26:56.652809] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # return 0 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.606 [2024-10-17 17:26:57.387506] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.606 [2024-10-17 17:26:57.411784] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.606 NULL1 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.606 Delay0 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=4037251 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:05:49.606 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:05:49.867 [2024-10-17 17:26:57.528793] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:05:51.784 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:05:51.784 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:51.784 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 starting I/O failed: -6 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 [2024-10-17 17:26:59.693235] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe65fd0 is same with the state(6) to be set 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Read completed with error (sct=0, sc=8) 00:05:51.784 Write completed with error (sct=0, sc=8) 00:05:51.785 [2024-10-17 17:26:59.694377] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe66390 is same with the state(6) to be set 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 Write completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 Read completed with error (sct=0, sc=8) 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:51.785 starting I/O failed: -6 00:05:53.170 [2024-10-17 17:27:00.666183] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe676b0 is same with the state(6) to be set 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 [2024-10-17 17:27:00.696596] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe661b0 is same with the state(6) to be set 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 [2024-10-17 17:27:00.697005] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe666c0 is same with the state(6) to be set 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 [2024-10-17 17:27:00.701398] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7fa47000d310 is same with the state(6) to be set 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Read completed with error (sct=0, sc=8) 00:05:53.170 Write completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Write completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Write completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 Read completed with error (sct=0, sc=8) 00:05:53.171 [2024-10-17 17:27:00.701884] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7fa470000c00 is same with the state(6) to be set 00:05:53.171 Initializing NVMe Controllers 00:05:53.171 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:05:53.171 Controller IO queue size 128, less than required. 00:05:53.171 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:05:53.171 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:05:53.171 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:05:53.171 Initialization complete. Launching workers. 00:05:53.171 ======================================================== 00:05:53.171 Latency(us) 00:05:53.171 Device Information : IOPS MiB/s Average min max 00:05:53.171 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 162.35 0.08 911339.68 1095.72 1006544.79 00:05:53.171 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 172.31 0.08 987375.88 382.76 2002781.55 00:05:53.171 ======================================================== 00:05:53.171 Total : 334.67 0.16 950489.27 382.76 2002781.55 00:05:53.171 00:05:53.171 [2024-10-17 17:27:00.702366] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe676b0 (9): Bad file descriptor 00:05:53.171 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:05:53.171 17:27:00 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:53.171 17:27:00 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:05:53.171 17:27:00 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 4037251 00:05:53.171 17:27:00 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 4037251 00:05:53.432 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (4037251) - No such process 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 4037251 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@650 -- # local es=0 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # valid_exec_arg wait 4037251 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@638 -- # local arg=wait 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # type -t wait 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # wait 4037251 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # es=1 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:05:53.432 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:53.433 [2024-10-17 17:27:01.231338] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=4038165 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:53.433 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:05:53.433 [2024-10-17 17:27:01.320776] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:05:54.005 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:05:54.005 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:54.005 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:05:54.594 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:05:54.594 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:54.594 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:05:54.855 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:05:54.855 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:54.855 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:05:55.425 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:05:55.425 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:55.425 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:05:55.998 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:05:55.998 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:55.998 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:05:56.570 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:05:56.570 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:56.570 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:05:56.570 Initializing NVMe Controllers 00:05:56.570 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:05:56.570 Controller IO queue size 128, less than required. 00:05:56.570 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:05:56.570 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:05:56.570 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:05:56.570 Initialization complete. Launching workers. 00:05:56.570 ======================================================== 00:05:56.570 Latency(us) 00:05:56.570 Device Information : IOPS MiB/s Average min max 00:05:56.570 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1002013.47 1000164.83 1005081.81 00:05:56.570 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1002899.75 1000201.25 1008245.58 00:05:56.570 ======================================================== 00:05:56.570 Total : 256.00 0.12 1002456.61 1000164.83 1008245.58 00:05:56.570 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 4038165 00:05:57.141 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (4038165) - No such process 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 4038165 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@514 -- # nvmfcleanup 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:05:57.141 rmmod nvme_tcp 00:05:57.141 rmmod nvme_fabrics 00:05:57.141 rmmod nvme_keyring 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@515 -- # '[' -n 4037213 ']' 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # killprocess 4037213 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@950 -- # '[' -z 4037213 ']' 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # kill -0 4037213 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # uname 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4037213 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4037213' 00:05:57.141 killing process with pid 4037213 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@969 -- # kill 4037213 00:05:57.141 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@974 -- # wait 4037213 00:05:57.141 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:05:57.141 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:05:57.141 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:05:57.141 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:05:57.141 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-save 00:05:57.142 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:05:57.142 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-restore 00:05:57.142 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:05:57.142 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:05:57.142 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:05:57.142 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:05:57.142 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:05:59.692 00:05:59.692 real 0m18.490s 00:05:59.692 user 0m30.942s 00:05:59.692 sys 0m6.834s 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:05:59.692 ************************************ 00:05:59.692 END TEST nvmf_delete_subsystem 00:05:59.692 ************************************ 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:05:59.692 ************************************ 00:05:59.692 START TEST nvmf_host_management 00:05:59.692 ************************************ 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:05:59.692 * Looking for test storage... 00:05:59.692 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1691 -- # lcov --version 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:05:59.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.692 --rc genhtml_branch_coverage=1 00:05:59.692 --rc genhtml_function_coverage=1 00:05:59.692 --rc genhtml_legend=1 00:05:59.692 --rc geninfo_all_blocks=1 00:05:59.692 --rc geninfo_unexecuted_blocks=1 00:05:59.692 00:05:59.692 ' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:05:59.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.692 --rc genhtml_branch_coverage=1 00:05:59.692 --rc genhtml_function_coverage=1 00:05:59.692 --rc genhtml_legend=1 00:05:59.692 --rc geninfo_all_blocks=1 00:05:59.692 --rc geninfo_unexecuted_blocks=1 00:05:59.692 00:05:59.692 ' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:05:59.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.692 --rc genhtml_branch_coverage=1 00:05:59.692 --rc genhtml_function_coverage=1 00:05:59.692 --rc genhtml_legend=1 00:05:59.692 --rc geninfo_all_blocks=1 00:05:59.692 --rc geninfo_unexecuted_blocks=1 00:05:59.692 00:05:59.692 ' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:05:59.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.692 --rc genhtml_branch_coverage=1 00:05:59.692 --rc genhtml_function_coverage=1 00:05:59.692 --rc genhtml_legend=1 00:05:59.692 --rc geninfo_all_blocks=1 00:05:59.692 --rc geninfo_unexecuted_blocks=1 00:05:59.692 00:05:59.692 ' 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:59.692 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:59.693 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@474 -- # prepare_net_devs 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@436 -- # local -g is_hw=no 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@438 -- # remove_spdk_ns 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:05:59.693 17:27:07 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:06:07.843 Found 0000:31:00.0 (0x8086 - 0x159b) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:06:07.843 Found 0000:31:00.1 (0x8086 - 0x159b) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:06:07.843 Found net devices under 0000:31:00.0: cvl_0_0 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:06:07.843 Found net devices under 0000:31:00.1: cvl_0_1 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # is_hw=yes 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:06:07.843 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:06:07.844 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:06:07.844 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.675 ms 00:06:07.844 00:06:07.844 --- 10.0.0.2 ping statistics --- 00:06:07.844 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:07.844 rtt min/avg/max/mdev = 0.675/0.675/0.675/0.000 ms 00:06:07.844 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:06:07.844 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:06:07.844 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.325 ms 00:06:07.844 00:06:07.844 --- 10.0.0.1 ping statistics --- 00:06:07.844 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:07.844 rtt min/avg/max/mdev = 0.325/0.325/0.325/0.000 ms 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@448 -- # return 0 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@507 -- # nvmfpid=4043697 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@508 -- # waitforlisten 4043697 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 4043697 ']' 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:07.844 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:07.844 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:07.844 [2024-10-17 17:27:15.125973] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:06:07.844 [2024-10-17 17:27:15.126045] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:07.844 [2024-10-17 17:27:15.216090] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:07.844 [2024-10-17 17:27:15.270596] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:06:07.844 [2024-10-17 17:27:15.270651] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:06:07.844 [2024-10-17 17:27:15.270660] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:07.844 [2024-10-17 17:27:15.270667] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:07.844 [2024-10-17 17:27:15.270673] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:06:07.844 [2024-10-17 17:27:15.273130] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:07.844 [2024-10-17 17:27:15.273288] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:07.844 [2024-10-17 17:27:15.273449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:06:07.844 [2024-10-17 17:27:15.273450] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:08.145 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:08.145 [2024-10-17 17:27:16.003115] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:08.145 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:08.440 Malloc0 00:06:08.440 [2024-10-17 17:27:16.083503] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=4043929 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 4043929 /var/tmp/bdevperf.sock 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 4043929 ']' 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:06:08.440 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:06:08.440 { 00:06:08.440 "params": { 00:06:08.440 "name": "Nvme$subsystem", 00:06:08.440 "trtype": "$TEST_TRANSPORT", 00:06:08.440 "traddr": "$NVMF_FIRST_TARGET_IP", 00:06:08.440 "adrfam": "ipv4", 00:06:08.440 "trsvcid": "$NVMF_PORT", 00:06:08.440 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:06:08.440 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:06:08.440 "hdgst": ${hdgst:-false}, 00:06:08.440 "ddgst": ${ddgst:-false} 00:06:08.440 }, 00:06:08.440 "method": "bdev_nvme_attach_controller" 00:06:08.440 } 00:06:08.440 EOF 00:06:08.440 )") 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:06:08.440 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:06:08.440 "params": { 00:06:08.440 "name": "Nvme0", 00:06:08.440 "trtype": "tcp", 00:06:08.440 "traddr": "10.0.0.2", 00:06:08.440 "adrfam": "ipv4", 00:06:08.440 "trsvcid": "4420", 00:06:08.440 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:06:08.440 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:06:08.440 "hdgst": false, 00:06:08.440 "ddgst": false 00:06:08.440 }, 00:06:08.440 "method": "bdev_nvme_attach_controller" 00:06:08.440 }' 00:06:08.440 [2024-10-17 17:27:16.192732] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:06:08.440 [2024-10-17 17:27:16.192800] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4043929 ] 00:06:08.440 [2024-10-17 17:27:16.277799] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:08.440 [2024-10-17 17:27:16.331327] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:09.019 Running I/O for 10 seconds... 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=643 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@58 -- # '[' 643 -ge 100 ']' 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@60 -- # break 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:09.283 [2024-10-17 17:27:17.167525] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167628] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167637] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167645] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167652] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167666] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167673] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167680] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167687] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167700] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167706] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167713] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167720] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167727] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167733] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167740] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167747] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167753] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167760] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167775] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167782] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167792] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 [2024-10-17 17:27:17.167799] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x6a1730 is same with the state(6) to be set 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:09.283 [2024-10-17 17:27:17.177061] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:06:09.283 [2024-10-17 17:27:17.177118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.177131] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:06:09.283 [2024-10-17 17:27:17.177139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.177148] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:06:09.283 [2024-10-17 17:27:17.177156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.177164] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:06:09.283 [2024-10-17 17:27:17.177173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.177180] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1f45c40 is same with the state(6) to be set 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:09.283 [2024-10-17 17:27:17.184794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:98048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.184854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:98176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.184872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:98304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.184890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:98432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 17:27:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:06:09.283 [2024-10-17 17:27:17.184908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:98560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.184943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:98688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.184961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:98816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.184980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:98944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.184988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.283 [2024-10-17 17:27:17.184997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:99072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.283 [2024-10-17 17:27:17.185005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:99200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:99328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:99456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:99584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:99712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:99840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:99968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:100096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:100224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:100352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:100480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:100608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:100736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:100864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:100992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:101120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:101248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:101376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:101504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:101632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:101760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:101888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185413] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:102016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:102144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:102272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:102400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:102528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:102656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:102784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:102912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:103040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:103168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:103296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:103424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:103552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:103680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:103808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.284 [2024-10-17 17:27:17.185710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:103936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.284 [2024-10-17 17:27:17.185718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:104064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:104192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:104320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:104448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:104576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:104704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:104832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:104960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:105088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:105216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:105344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:105472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:105600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:105728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.185983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.185993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:105856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.186001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.186010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:105984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.186019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.186029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:106112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:06:09.285 [2024-10-17 17:27:17.186038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:06:09.285 [2024-10-17 17:27:17.186121] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1f56570 was disconnected and freed. reset controller. 00:06:09.285 [2024-10-17 17:27:17.187329] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:06:09.285 task offset: 98048 on job bdev=Nvme0n1 fails 00:06:09.285 00:06:09.285 Latency(us) 00:06:09.285 [2024-10-17T15:27:17.204Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:06:09.285 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:06:09.285 Job: Nvme0n1 ended in about 0.48 seconds with error 00:06:09.285 Verification LBA range: start 0x0 length 0x400 00:06:09.285 Nvme0n1 : 0.48 1585.55 99.10 132.47 0.00 36217.37 1720.32 34734.08 00:06:09.285 [2024-10-17T15:27:17.204Z] =================================================================================================================== 00:06:09.285 [2024-10-17T15:27:17.204Z] Total : 1585.55 99.10 132.47 0.00 36217.37 1720.32 34734.08 00:06:09.285 [2024-10-17 17:27:17.189551] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:09.285 [2024-10-17 17:27:17.189590] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1f45c40 (9): Bad file descriptor 00:06:09.544 [2024-10-17 17:27:17.201514] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 4043929 00:06:10.486 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (4043929) - No such process 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # true 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:06:10.486 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:06:10.486 { 00:06:10.486 "params": { 00:06:10.486 "name": "Nvme$subsystem", 00:06:10.486 "trtype": "$TEST_TRANSPORT", 00:06:10.486 "traddr": "$NVMF_FIRST_TARGET_IP", 00:06:10.486 "adrfam": "ipv4", 00:06:10.486 "trsvcid": "$NVMF_PORT", 00:06:10.486 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:06:10.486 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:06:10.486 "hdgst": ${hdgst:-false}, 00:06:10.486 "ddgst": ${ddgst:-false} 00:06:10.487 }, 00:06:10.487 "method": "bdev_nvme_attach_controller" 00:06:10.487 } 00:06:10.487 EOF 00:06:10.487 )") 00:06:10.487 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:06:10.487 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:06:10.487 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:06:10.487 17:27:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:06:10.487 "params": { 00:06:10.487 "name": "Nvme0", 00:06:10.487 "trtype": "tcp", 00:06:10.487 "traddr": "10.0.0.2", 00:06:10.487 "adrfam": "ipv4", 00:06:10.487 "trsvcid": "4420", 00:06:10.487 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:06:10.487 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:06:10.487 "hdgst": false, 00:06:10.487 "ddgst": false 00:06:10.487 }, 00:06:10.487 "method": "bdev_nvme_attach_controller" 00:06:10.487 }' 00:06:10.487 [2024-10-17 17:27:18.244179] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:06:10.487 [2024-10-17 17:27:18.244233] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4044291 ] 00:06:10.487 [2024-10-17 17:27:18.324510] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:10.487 [2024-10-17 17:27:18.361407] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:10.747 Running I/O for 1 seconds... 00:06:11.687 1604.00 IOPS, 100.25 MiB/s 00:06:11.687 Latency(us) 00:06:11.687 [2024-10-17T15:27:19.606Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:06:11.687 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:06:11.687 Verification LBA range: start 0x0 length 0x400 00:06:11.687 Nvme0n1 : 1.02 1639.47 102.47 0.00 0.00 38342.77 5980.16 32768.00 00:06:11.687 [2024-10-17T15:27:19.606Z] =================================================================================================================== 00:06:11.687 [2024-10-17T15:27:19.606Z] Total : 1639.47 102.47 0.00 0.00 38342.77 5980.16 32768.00 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@514 -- # nvmfcleanup 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:06:11.947 rmmod nvme_tcp 00:06:11.947 rmmod nvme_fabrics 00:06:11.947 rmmod nvme_keyring 00:06:11.947 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@515 -- # '[' -n 4043697 ']' 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@516 -- # killprocess 4043697 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@950 -- # '[' -z 4043697 ']' 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@954 -- # kill -0 4043697 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@955 -- # uname 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4043697 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4043697' 00:06:11.948 killing process with pid 4043697 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@969 -- # kill 4043697 00:06:11.948 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@974 -- # wait 4043697 00:06:12.209 [2024-10-17 17:27:19.881459] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-save 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-restore 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:12.209 17:27:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:14.197 17:27:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:06:14.197 17:27:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:06:14.197 00:06:14.197 real 0m14.804s 00:06:14.197 user 0m23.307s 00:06:14.197 sys 0m6.853s 00:06:14.197 17:27:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:14.197 17:27:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:06:14.197 ************************************ 00:06:14.197 END TEST nvmf_host_management 00:06:14.197 ************************************ 00:06:14.197 17:27:22 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:06:14.197 17:27:22 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:14.197 17:27:22 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:14.197 17:27:22 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:06:14.197 ************************************ 00:06:14.197 START TEST nvmf_lvol 00:06:14.197 ************************************ 00:06:14.198 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:06:14.460 * Looking for test storage... 00:06:14.460 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1691 -- # lcov --version 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:14.460 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:14.460 --rc genhtml_branch_coverage=1 00:06:14.460 --rc genhtml_function_coverage=1 00:06:14.460 --rc genhtml_legend=1 00:06:14.460 --rc geninfo_all_blocks=1 00:06:14.460 --rc geninfo_unexecuted_blocks=1 00:06:14.460 00:06:14.460 ' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:14.460 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:14.460 --rc genhtml_branch_coverage=1 00:06:14.460 --rc genhtml_function_coverage=1 00:06:14.460 --rc genhtml_legend=1 00:06:14.460 --rc geninfo_all_blocks=1 00:06:14.460 --rc geninfo_unexecuted_blocks=1 00:06:14.460 00:06:14.460 ' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:14.460 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:14.460 --rc genhtml_branch_coverage=1 00:06:14.460 --rc genhtml_function_coverage=1 00:06:14.460 --rc genhtml_legend=1 00:06:14.460 --rc geninfo_all_blocks=1 00:06:14.460 --rc geninfo_unexecuted_blocks=1 00:06:14.460 00:06:14.460 ' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:14.460 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:14.460 --rc genhtml_branch_coverage=1 00:06:14.460 --rc genhtml_function_coverage=1 00:06:14.460 --rc genhtml_legend=1 00:06:14.460 --rc geninfo_all_blocks=1 00:06:14.460 --rc geninfo_unexecuted_blocks=1 00:06:14.460 00:06:14.460 ' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:14.460 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:14.461 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@474 -- # prepare_net_devs 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@436 -- # local -g is_hw=no 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@438 -- # remove_spdk_ns 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:06:14.461 17:27:22 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:06:22.600 Found 0000:31:00.0 (0x8086 - 0x159b) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:06:22.600 Found 0000:31:00.1 (0x8086 - 0x159b) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:06:22.600 Found net devices under 0000:31:00.0: cvl_0_0 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:06:22.600 Found net devices under 0000:31:00.1: cvl_0_1 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:06:22.600 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # is_hw=yes 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:06:22.601 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:06:22.601 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.647 ms 00:06:22.601 00:06:22.601 --- 10.0.0.2 ping statistics --- 00:06:22.601 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:22.601 rtt min/avg/max/mdev = 0.647/0.647/0.647/0.000 ms 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:06:22.601 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:06:22.601 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.289 ms 00:06:22.601 00:06:22.601 --- 10.0.0.1 ping statistics --- 00:06:22.601 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:22.601 rtt min/avg/max/mdev = 0.289/0.289/0.289/0.000 ms 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@448 -- # return 0 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@724 -- # xtrace_disable 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@507 -- # nvmfpid=4048999 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@508 -- # waitforlisten 4048999 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@831 -- # '[' -z 4048999 ']' 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:22.601 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:22.601 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:06:22.601 [2024-10-17 17:27:30.054291] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:06:22.601 [2024-10-17 17:27:30.054369] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:22.601 [2024-10-17 17:27:30.143793] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:22.601 [2024-10-17 17:27:30.197091] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:06:22.601 [2024-10-17 17:27:30.197148] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:06:22.601 [2024-10-17 17:27:30.197157] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:22.601 [2024-10-17 17:27:30.197165] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:22.601 [2024-10-17 17:27:30.197176] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:06:22.601 [2024-10-17 17:27:30.199066] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:22.601 [2024-10-17 17:27:30.199225] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:22.601 [2024-10-17 17:27:30.199226] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:23.171 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:23.171 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@864 -- # return 0 00:06:23.171 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:06:23.171 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@730 -- # xtrace_disable 00:06:23.171 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:06:23.171 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:06:23.171 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:06:23.432 [2024-10-17 17:27:31.091365] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:23.432 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:06:23.692 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:06:23.692 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:06:23.692 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:06:23.692 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:06:23.953 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:06:24.213 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=60f61323-f22f-4d50-a18f-be560daafada 00:06:24.213 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 60f61323-f22f-4d50-a18f-be560daafada lvol 20 00:06:24.475 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=c77f7f0e-b294-47fe-9458-f8c0ceac385e 00:06:24.475 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:06:24.475 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 c77f7f0e-b294-47fe-9458-f8c0ceac385e 00:06:24.736 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:06:24.996 [2024-10-17 17:27:32.698203] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:24.996 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:06:24.996 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=4049697 00:06:24.996 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:06:24.996 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:06:26.377 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot c77f7f0e-b294-47fe-9458-f8c0ceac385e MY_SNAPSHOT 00:06:26.377 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=53eca4d4-166f-4ff9-88ad-941c752197e4 00:06:26.377 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize c77f7f0e-b294-47fe-9458-f8c0ceac385e 30 00:06:26.637 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone 53eca4d4-166f-4ff9-88ad-941c752197e4 MY_CLONE 00:06:26.637 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=dfa3ff63-ea99-4926-929f-43dd7a2bfda1 00:06:26.637 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate dfa3ff63-ea99-4926-929f-43dd7a2bfda1 00:06:27.207 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 4049697 00:06:35.341 Initializing NVMe Controllers 00:06:35.341 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:06:35.341 Controller IO queue size 128, less than required. 00:06:35.341 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:06:35.341 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:06:35.341 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:06:35.341 Initialization complete. Launching workers. 00:06:35.341 ======================================================== 00:06:35.341 Latency(us) 00:06:35.341 Device Information : IOPS MiB/s Average min max 00:06:35.341 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 16670.80 65.12 7679.64 1481.10 41654.08 00:06:35.341 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 17294.20 67.56 7402.13 1281.62 39604.19 00:06:35.341 ======================================================== 00:06:35.342 Total : 33965.00 132.68 7538.34 1281.62 41654.08 00:06:35.342 00:06:35.342 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:06:35.601 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete c77f7f0e-b294-47fe-9458-f8c0ceac385e 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 60f61323-f22f-4d50-a18f-be560daafada 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@514 -- # nvmfcleanup 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:06:35.862 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:06:35.862 rmmod nvme_tcp 00:06:36.122 rmmod nvme_fabrics 00:06:36.122 rmmod nvme_keyring 00:06:36.122 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:06:36.122 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:06:36.122 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:06:36.122 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@515 -- # '[' -n 4048999 ']' 00:06:36.122 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@516 -- # killprocess 4048999 00:06:36.122 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@950 -- # '[' -z 4048999 ']' 00:06:36.122 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@954 -- # kill -0 4048999 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@955 -- # uname 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4048999 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4048999' 00:06:36.123 killing process with pid 4048999 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@969 -- # kill 4048999 00:06:36.123 17:27:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@974 -- # wait 4048999 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-save 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-restore 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:36.123 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:06:38.665 00:06:38.665 real 0m24.022s 00:06:38.665 user 1m4.633s 00:06:38.665 sys 0m8.657s 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:06:38.665 ************************************ 00:06:38.665 END TEST nvmf_lvol 00:06:38.665 ************************************ 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:06:38.665 ************************************ 00:06:38.665 START TEST nvmf_lvs_grow 00:06:38.665 ************************************ 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:06:38.665 * Looking for test storage... 00:06:38.665 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1691 -- # lcov --version 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:38.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.665 --rc genhtml_branch_coverage=1 00:06:38.665 --rc genhtml_function_coverage=1 00:06:38.665 --rc genhtml_legend=1 00:06:38.665 --rc geninfo_all_blocks=1 00:06:38.665 --rc geninfo_unexecuted_blocks=1 00:06:38.665 00:06:38.665 ' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:38.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.665 --rc genhtml_branch_coverage=1 00:06:38.665 --rc genhtml_function_coverage=1 00:06:38.665 --rc genhtml_legend=1 00:06:38.665 --rc geninfo_all_blocks=1 00:06:38.665 --rc geninfo_unexecuted_blocks=1 00:06:38.665 00:06:38.665 ' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:38.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.665 --rc genhtml_branch_coverage=1 00:06:38.665 --rc genhtml_function_coverage=1 00:06:38.665 --rc genhtml_legend=1 00:06:38.665 --rc geninfo_all_blocks=1 00:06:38.665 --rc geninfo_unexecuted_blocks=1 00:06:38.665 00:06:38.665 ' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:38.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.665 --rc genhtml_branch_coverage=1 00:06:38.665 --rc genhtml_function_coverage=1 00:06:38.665 --rc genhtml_legend=1 00:06:38.665 --rc geninfo_all_blocks=1 00:06:38.665 --rc geninfo_unexecuted_blocks=1 00:06:38.665 00:06:38.665 ' 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:38.665 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:38.666 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@474 -- # prepare_net_devs 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@436 -- # local -g is_hw=no 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@438 -- # remove_spdk_ns 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:06:38.666 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:06:46.804 Found 0000:31:00.0 (0x8086 - 0x159b) 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:46.804 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:06:46.805 Found 0000:31:00.1 (0x8086 - 0x159b) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:06:46.805 Found net devices under 0000:31:00.0: cvl_0_0 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:06:46.805 Found net devices under 0000:31:00.1: cvl_0_1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # is_hw=yes 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:06:46.805 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:06:46.805 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.649 ms 00:06:46.805 00:06:46.805 --- 10.0.0.2 ping statistics --- 00:06:46.805 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:46.805 rtt min/avg/max/mdev = 0.649/0.649/0.649/0.000 ms 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:06:46.805 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:06:46.805 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.284 ms 00:06:46.805 00:06:46.805 --- 10.0.0.1 ping statistics --- 00:06:46.805 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:46.805 rtt min/avg/max/mdev = 0.284/0.284/0.284/0.000 ms 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@448 -- # return 0 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@724 -- # xtrace_disable 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@507 -- # nvmfpid=4056106 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@508 -- # waitforlisten 4056106 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@831 -- # '[' -z 4056106 ']' 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:46.805 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:46.805 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:06:46.805 [2024-10-17 17:27:54.023570] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:06:46.805 [2024-10-17 17:27:54.023635] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:46.805 [2024-10-17 17:27:54.111184] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:46.805 [2024-10-17 17:27:54.162005] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:06:46.805 [2024-10-17 17:27:54.162053] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:06:46.805 [2024-10-17 17:27:54.162062] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:46.805 [2024-10-17 17:27:54.162069] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:46.805 [2024-10-17 17:27:54.162076] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:06:46.805 [2024-10-17 17:27:54.162898] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:47.067 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:47.067 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # return 0 00:06:47.067 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:06:47.067 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@730 -- # xtrace_disable 00:06:47.067 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:06:47.067 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:06:47.067 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:06:47.328 [2024-10-17 17:27:55.057328] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:06:47.328 ************************************ 00:06:47.328 START TEST lvs_grow_clean 00:06:47.328 ************************************ 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1125 -- # lvs_grow 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:06:47.328 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:06:47.589 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:06:47.589 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:06:47.849 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:06:47.849 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:06:47.849 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:06:47.849 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:06:47.849 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:06:47.849 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 lvol 150 00:06:48.110 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=083fe500-ce80-494c-8321-be0e294433e5 00:06:48.110 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:06:48.110 17:27:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:06:48.370 [2024-10-17 17:27:56.107582] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:06:48.370 [2024-10-17 17:27:56.107651] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:06:48.370 true 00:06:48.370 17:27:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:06:48.370 17:27:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:06:48.631 17:27:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:06:48.631 17:27:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:06:48.631 17:27:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 083fe500-ce80-494c-8321-be0e294433e5 00:06:48.892 17:27:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:06:49.152 [2024-10-17 17:27:56.829902] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:49.152 17:27:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=4056815 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 4056815 /var/tmp/bdevperf.sock 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@831 -- # '[' -z 4056815 ']' 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:06:49.152 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:49.152 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:06:49.412 [2024-10-17 17:27:57.085744] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:06:49.412 [2024-10-17 17:27:57.085821] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4056815 ] 00:06:49.412 [2024-10-17 17:27:57.176712] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:49.412 [2024-10-17 17:27:57.228214] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:50.351 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:50.351 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # return 0 00:06:50.351 17:27:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:06:50.351 Nvme0n1 00:06:50.351 17:27:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:06:50.612 [ 00:06:50.612 { 00:06:50.612 "name": "Nvme0n1", 00:06:50.612 "aliases": [ 00:06:50.612 "083fe500-ce80-494c-8321-be0e294433e5" 00:06:50.612 ], 00:06:50.612 "product_name": "NVMe disk", 00:06:50.612 "block_size": 4096, 00:06:50.612 "num_blocks": 38912, 00:06:50.612 "uuid": "083fe500-ce80-494c-8321-be0e294433e5", 00:06:50.612 "numa_id": 0, 00:06:50.612 "assigned_rate_limits": { 00:06:50.612 "rw_ios_per_sec": 0, 00:06:50.612 "rw_mbytes_per_sec": 0, 00:06:50.612 "r_mbytes_per_sec": 0, 00:06:50.612 "w_mbytes_per_sec": 0 00:06:50.612 }, 00:06:50.612 "claimed": false, 00:06:50.612 "zoned": false, 00:06:50.612 "supported_io_types": { 00:06:50.612 "read": true, 00:06:50.612 "write": true, 00:06:50.612 "unmap": true, 00:06:50.612 "flush": true, 00:06:50.612 "reset": true, 00:06:50.612 "nvme_admin": true, 00:06:50.612 "nvme_io": true, 00:06:50.612 "nvme_io_md": false, 00:06:50.612 "write_zeroes": true, 00:06:50.612 "zcopy": false, 00:06:50.612 "get_zone_info": false, 00:06:50.612 "zone_management": false, 00:06:50.612 "zone_append": false, 00:06:50.612 "compare": true, 00:06:50.612 "compare_and_write": true, 00:06:50.612 "abort": true, 00:06:50.612 "seek_hole": false, 00:06:50.612 "seek_data": false, 00:06:50.612 "copy": true, 00:06:50.612 "nvme_iov_md": false 00:06:50.612 }, 00:06:50.612 "memory_domains": [ 00:06:50.612 { 00:06:50.612 "dma_device_id": "system", 00:06:50.612 "dma_device_type": 1 00:06:50.612 } 00:06:50.612 ], 00:06:50.612 "driver_specific": { 00:06:50.612 "nvme": [ 00:06:50.612 { 00:06:50.612 "trid": { 00:06:50.612 "trtype": "TCP", 00:06:50.612 "adrfam": "IPv4", 00:06:50.612 "traddr": "10.0.0.2", 00:06:50.612 "trsvcid": "4420", 00:06:50.612 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:06:50.612 }, 00:06:50.612 "ctrlr_data": { 00:06:50.612 "cntlid": 1, 00:06:50.612 "vendor_id": "0x8086", 00:06:50.612 "model_number": "SPDK bdev Controller", 00:06:50.612 "serial_number": "SPDK0", 00:06:50.612 "firmware_revision": "25.01", 00:06:50.612 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:06:50.612 "oacs": { 00:06:50.612 "security": 0, 00:06:50.612 "format": 0, 00:06:50.612 "firmware": 0, 00:06:50.612 "ns_manage": 0 00:06:50.612 }, 00:06:50.612 "multi_ctrlr": true, 00:06:50.612 "ana_reporting": false 00:06:50.612 }, 00:06:50.612 "vs": { 00:06:50.612 "nvme_version": "1.3" 00:06:50.612 }, 00:06:50.612 "ns_data": { 00:06:50.612 "id": 1, 00:06:50.612 "can_share": true 00:06:50.612 } 00:06:50.612 } 00:06:50.612 ], 00:06:50.612 "mp_policy": "active_passive" 00:06:50.612 } 00:06:50.612 } 00:06:50.612 ] 00:06:50.612 17:27:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:06:50.612 17:27:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=4057014 00:06:50.612 17:27:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:06:50.612 Running I/O for 10 seconds... 00:06:51.554 Latency(us) 00:06:51.554 [2024-10-17T15:27:59.473Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:06:51.554 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:51.554 Nvme0n1 : 1.00 25169.00 98.32 0.00 0.00 0.00 0.00 0.00 00:06:51.554 [2024-10-17T15:27:59.473Z] =================================================================================================================== 00:06:51.554 [2024-10-17T15:27:59.473Z] Total : 25169.00 98.32 0.00 0.00 0.00 0.00 0.00 00:06:51.554 00:06:52.498 17:28:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:06:52.757 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:52.757 Nvme0n1 : 2.00 25301.00 98.83 0.00 0.00 0.00 0.00 0.00 00:06:52.757 [2024-10-17T15:28:00.676Z] =================================================================================================================== 00:06:52.757 [2024-10-17T15:28:00.676Z] Total : 25301.00 98.83 0.00 0.00 0.00 0.00 0.00 00:06:52.757 00:06:52.757 true 00:06:52.757 17:28:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:06:52.757 17:28:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:06:53.016 17:28:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:06:53.016 17:28:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:06:53.016 17:28:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 4057014 00:06:53.585 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:53.585 Nvme0n1 : 3.00 25369.00 99.10 0.00 0.00 0.00 0.00 0.00 00:06:53.585 [2024-10-17T15:28:01.504Z] =================================================================================================================== 00:06:53.585 [2024-10-17T15:28:01.504Z] Total : 25369.00 99.10 0.00 0.00 0.00 0.00 0.00 00:06:53.585 00:06:54.965 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:54.965 Nvme0n1 : 4.00 25428.50 99.33 0.00 0.00 0.00 0.00 0.00 00:06:54.965 [2024-10-17T15:28:02.884Z] =================================================================================================================== 00:06:54.965 [2024-10-17T15:28:02.884Z] Total : 25428.50 99.33 0.00 0.00 0.00 0.00 0.00 00:06:54.965 00:06:55.905 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:55.905 Nvme0n1 : 5.00 25461.60 99.46 0.00 0.00 0.00 0.00 0.00 00:06:55.905 [2024-10-17T15:28:03.824Z] =================================================================================================================== 00:06:55.905 [2024-10-17T15:28:03.824Z] Total : 25461.60 99.46 0.00 0.00 0.00 0.00 0.00 00:06:55.905 00:06:56.846 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:56.846 Nvme0n1 : 6.00 25495.17 99.59 0.00 0.00 0.00 0.00 0.00 00:06:56.846 [2024-10-17T15:28:04.765Z] =================================================================================================================== 00:06:56.846 [2024-10-17T15:28:04.765Z] Total : 25495.17 99.59 0.00 0.00 0.00 0.00 0.00 00:06:56.846 00:06:57.787 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:57.787 Nvme0n1 : 7.00 25509.57 99.65 0.00 0.00 0.00 0.00 0.00 00:06:57.787 [2024-10-17T15:28:05.706Z] =================================================================================================================== 00:06:57.787 [2024-10-17T15:28:05.706Z] Total : 25509.57 99.65 0.00 0.00 0.00 0.00 0.00 00:06:57.787 00:06:58.727 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:58.727 Nvme0n1 : 8.00 25528.62 99.72 0.00 0.00 0.00 0.00 0.00 00:06:58.727 [2024-10-17T15:28:06.646Z] =================================================================================================================== 00:06:58.727 [2024-10-17T15:28:06.646Z] Total : 25528.62 99.72 0.00 0.00 0.00 0.00 0.00 00:06:58.727 00:06:59.667 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:06:59.667 Nvme0n1 : 9.00 25543.44 99.78 0.00 0.00 0.00 0.00 0.00 00:06:59.667 [2024-10-17T15:28:07.586Z] =================================================================================================================== 00:06:59.667 [2024-10-17T15:28:07.586Z] Total : 25543.44 99.78 0.00 0.00 0.00 0.00 0.00 00:06:59.667 00:07:00.609 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:00.609 Nvme0n1 : 10.00 25555.60 99.83 0.00 0.00 0.00 0.00 0.00 00:07:00.609 [2024-10-17T15:28:08.528Z] =================================================================================================================== 00:07:00.609 [2024-10-17T15:28:08.528Z] Total : 25555.60 99.83 0.00 0.00 0.00 0.00 0.00 00:07:00.609 00:07:00.609 00:07:00.609 Latency(us) 00:07:00.609 [2024-10-17T15:28:08.528Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:00.609 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:00.609 Nvme0n1 : 10.00 25552.72 99.82 0.00 0.00 5005.68 2143.57 8628.91 00:07:00.609 [2024-10-17T15:28:08.528Z] =================================================================================================================== 00:07:00.609 [2024-10-17T15:28:08.528Z] Total : 25552.72 99.82 0.00 0.00 5005.68 2143.57 8628.91 00:07:00.609 { 00:07:00.609 "results": [ 00:07:00.609 { 00:07:00.609 "job": "Nvme0n1", 00:07:00.609 "core_mask": "0x2", 00:07:00.609 "workload": "randwrite", 00:07:00.609 "status": "finished", 00:07:00.609 "queue_depth": 128, 00:07:00.609 "io_size": 4096, 00:07:00.609 "runtime": 10.003591, 00:07:00.609 "iops": 25552.724016805565, 00:07:00.609 "mibps": 99.81532819064674, 00:07:00.609 "io_failed": 0, 00:07:00.609 "io_timeout": 0, 00:07:00.609 "avg_latency_us": 5005.679084992378, 00:07:00.609 "min_latency_us": 2143.5733333333333, 00:07:00.609 "max_latency_us": 8628.906666666666 00:07:00.609 } 00:07:00.609 ], 00:07:00.609 "core_count": 1 00:07:00.609 } 00:07:00.609 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 4056815 00:07:00.609 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@950 -- # '[' -z 4056815 ']' 00:07:00.609 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # kill -0 4056815 00:07:00.609 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # uname 00:07:00.609 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:00.609 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4056815 00:07:00.869 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:07:00.869 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:07:00.869 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4056815' 00:07:00.869 killing process with pid 4056815 00:07:00.869 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@969 -- # kill 4056815 00:07:00.869 Received shutdown signal, test time was about 10.000000 seconds 00:07:00.869 00:07:00.869 Latency(us) 00:07:00.869 [2024-10-17T15:28:08.788Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:00.869 [2024-10-17T15:28:08.788Z] =================================================================================================================== 00:07:00.869 [2024-10-17T15:28:08.788Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:07:00.869 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@974 -- # wait 4056815 00:07:00.869 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:01.129 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:07:01.129 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:07:01.129 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:07:01.388 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:07:01.388 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:07:01.388 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:07:01.648 [2024-10-17 17:28:09.349966] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@650 -- # local es=0 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:07:01.648 request: 00:07:01.648 { 00:07:01.648 "uuid": "6989d6e4-bc11-4139-9d0b-ef6ac16f3767", 00:07:01.648 "method": "bdev_lvol_get_lvstores", 00:07:01.648 "req_id": 1 00:07:01.648 } 00:07:01.648 Got JSON-RPC error response 00:07:01.648 response: 00:07:01.648 { 00:07:01.648 "code": -19, 00:07:01.648 "message": "No such device" 00:07:01.648 } 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # es=1 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:01.648 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:07:01.909 aio_bdev 00:07:01.909 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 083fe500-ce80-494c-8321-be0e294433e5 00:07:01.909 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@899 -- # local bdev_name=083fe500-ce80-494c-8321-be0e294433e5 00:07:01.909 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:01.909 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@901 -- # local i 00:07:01.909 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:01.909 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:01.909 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:07:02.170 17:28:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 083fe500-ce80-494c-8321-be0e294433e5 -t 2000 00:07:02.170 [ 00:07:02.170 { 00:07:02.170 "name": "083fe500-ce80-494c-8321-be0e294433e5", 00:07:02.170 "aliases": [ 00:07:02.170 "lvs/lvol" 00:07:02.170 ], 00:07:02.170 "product_name": "Logical Volume", 00:07:02.170 "block_size": 4096, 00:07:02.170 "num_blocks": 38912, 00:07:02.170 "uuid": "083fe500-ce80-494c-8321-be0e294433e5", 00:07:02.170 "assigned_rate_limits": { 00:07:02.170 "rw_ios_per_sec": 0, 00:07:02.170 "rw_mbytes_per_sec": 0, 00:07:02.170 "r_mbytes_per_sec": 0, 00:07:02.170 "w_mbytes_per_sec": 0 00:07:02.170 }, 00:07:02.170 "claimed": false, 00:07:02.170 "zoned": false, 00:07:02.170 "supported_io_types": { 00:07:02.170 "read": true, 00:07:02.170 "write": true, 00:07:02.170 "unmap": true, 00:07:02.170 "flush": false, 00:07:02.170 "reset": true, 00:07:02.170 "nvme_admin": false, 00:07:02.170 "nvme_io": false, 00:07:02.170 "nvme_io_md": false, 00:07:02.170 "write_zeroes": true, 00:07:02.170 "zcopy": false, 00:07:02.170 "get_zone_info": false, 00:07:02.170 "zone_management": false, 00:07:02.170 "zone_append": false, 00:07:02.170 "compare": false, 00:07:02.170 "compare_and_write": false, 00:07:02.170 "abort": false, 00:07:02.170 "seek_hole": true, 00:07:02.170 "seek_data": true, 00:07:02.170 "copy": false, 00:07:02.170 "nvme_iov_md": false 00:07:02.170 }, 00:07:02.170 "driver_specific": { 00:07:02.170 "lvol": { 00:07:02.170 "lvol_store_uuid": "6989d6e4-bc11-4139-9d0b-ef6ac16f3767", 00:07:02.170 "base_bdev": "aio_bdev", 00:07:02.170 "thin_provision": false, 00:07:02.170 "num_allocated_clusters": 38, 00:07:02.170 "snapshot": false, 00:07:02.170 "clone": false, 00:07:02.170 "esnap_clone": false 00:07:02.170 } 00:07:02.170 } 00:07:02.170 } 00:07:02.170 ] 00:07:02.170 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@907 -- # return 0 00:07:02.170 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:07:02.170 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:07:02.430 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:07:02.430 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:07:02.430 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:07:02.690 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:07:02.690 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 083fe500-ce80-494c-8321-be0e294433e5 00:07:02.690 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 6989d6e4-bc11-4139-9d0b-ef6ac16f3767 00:07:02.950 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:07:03.211 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:03.211 00:07:03.211 real 0m15.816s 00:07:03.211 user 0m15.584s 00:07:03.211 sys 0m1.389s 00:07:03.211 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:03.211 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:07:03.211 ************************************ 00:07:03.211 END TEST lvs_grow_clean 00:07:03.211 ************************************ 00:07:03.211 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:07:03.211 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:03.211 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:03.211 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:03.211 ************************************ 00:07:03.211 START TEST lvs_grow_dirty 00:07:03.211 ************************************ 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1125 -- # lvs_grow dirty 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:03.211 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:07:03.472 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:07:03.472 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:07:03.732 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:03.732 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:03.732 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:07:03.732 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:07:03.732 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:07:03.732 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 4d250600-c9ba-44c0-89d7-250e049e2a39 lvol 150 00:07:03.992 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=a4a89185-6e85-47b5-928c-08e39fe65533 00:07:03.992 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:03.992 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:07:04.252 [2024-10-17 17:28:11.915399] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:07:04.252 [2024-10-17 17:28:11.915438] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:07:04.252 true 00:07:04.252 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:07:04.252 17:28:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:04.252 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:07:04.252 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:07:04.513 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 a4a89185-6e85-47b5-928c-08e39fe65533 00:07:04.773 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:07:04.773 [2024-10-17 17:28:12.589348] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:04.773 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=4059913 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 4059913 /var/tmp/bdevperf.sock 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 4059913 ']' 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:07:05.034 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:05.034 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:07:05.034 [2024-10-17 17:28:12.820218] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:05.034 [2024-10-17 17:28:12.820266] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4059913 ] 00:07:05.034 [2024-10-17 17:28:12.896772] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:05.034 [2024-10-17 17:28:12.926353] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:05.294 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:05.294 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:07:05.294 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:07:05.555 Nvme0n1 00:07:05.555 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:07:05.555 [ 00:07:05.555 { 00:07:05.555 "name": "Nvme0n1", 00:07:05.555 "aliases": [ 00:07:05.555 "a4a89185-6e85-47b5-928c-08e39fe65533" 00:07:05.555 ], 00:07:05.555 "product_name": "NVMe disk", 00:07:05.555 "block_size": 4096, 00:07:05.555 "num_blocks": 38912, 00:07:05.555 "uuid": "a4a89185-6e85-47b5-928c-08e39fe65533", 00:07:05.555 "numa_id": 0, 00:07:05.555 "assigned_rate_limits": { 00:07:05.555 "rw_ios_per_sec": 0, 00:07:05.555 "rw_mbytes_per_sec": 0, 00:07:05.555 "r_mbytes_per_sec": 0, 00:07:05.555 "w_mbytes_per_sec": 0 00:07:05.555 }, 00:07:05.555 "claimed": false, 00:07:05.555 "zoned": false, 00:07:05.555 "supported_io_types": { 00:07:05.555 "read": true, 00:07:05.555 "write": true, 00:07:05.555 "unmap": true, 00:07:05.555 "flush": true, 00:07:05.555 "reset": true, 00:07:05.555 "nvme_admin": true, 00:07:05.555 "nvme_io": true, 00:07:05.555 "nvme_io_md": false, 00:07:05.555 "write_zeroes": true, 00:07:05.555 "zcopy": false, 00:07:05.555 "get_zone_info": false, 00:07:05.555 "zone_management": false, 00:07:05.555 "zone_append": false, 00:07:05.555 "compare": true, 00:07:05.555 "compare_and_write": true, 00:07:05.555 "abort": true, 00:07:05.555 "seek_hole": false, 00:07:05.555 "seek_data": false, 00:07:05.555 "copy": true, 00:07:05.555 "nvme_iov_md": false 00:07:05.555 }, 00:07:05.555 "memory_domains": [ 00:07:05.555 { 00:07:05.555 "dma_device_id": "system", 00:07:05.555 "dma_device_type": 1 00:07:05.555 } 00:07:05.555 ], 00:07:05.555 "driver_specific": { 00:07:05.555 "nvme": [ 00:07:05.555 { 00:07:05.555 "trid": { 00:07:05.555 "trtype": "TCP", 00:07:05.555 "adrfam": "IPv4", 00:07:05.555 "traddr": "10.0.0.2", 00:07:05.555 "trsvcid": "4420", 00:07:05.555 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:07:05.555 }, 00:07:05.555 "ctrlr_data": { 00:07:05.555 "cntlid": 1, 00:07:05.555 "vendor_id": "0x8086", 00:07:05.555 "model_number": "SPDK bdev Controller", 00:07:05.555 "serial_number": "SPDK0", 00:07:05.555 "firmware_revision": "25.01", 00:07:05.555 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:07:05.555 "oacs": { 00:07:05.555 "security": 0, 00:07:05.555 "format": 0, 00:07:05.555 "firmware": 0, 00:07:05.555 "ns_manage": 0 00:07:05.555 }, 00:07:05.555 "multi_ctrlr": true, 00:07:05.555 "ana_reporting": false 00:07:05.555 }, 00:07:05.555 "vs": { 00:07:05.555 "nvme_version": "1.3" 00:07:05.555 }, 00:07:05.555 "ns_data": { 00:07:05.555 "id": 1, 00:07:05.555 "can_share": true 00:07:05.555 } 00:07:05.555 } 00:07:05.555 ], 00:07:05.555 "mp_policy": "active_passive" 00:07:05.555 } 00:07:05.555 } 00:07:05.555 ] 00:07:05.555 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=4059986 00:07:05.555 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:07:05.555 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:07:05.816 Running I/O for 10 seconds... 00:07:06.784 Latency(us) 00:07:06.784 [2024-10-17T15:28:14.703Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:06.784 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:06.784 Nvme0n1 : 1.00 25090.00 98.01 0.00 0.00 0.00 0.00 0.00 00:07:06.784 [2024-10-17T15:28:14.703Z] =================================================================================================================== 00:07:06.784 [2024-10-17T15:28:14.703Z] Total : 25090.00 98.01 0.00 0.00 0.00 0.00 0.00 00:07:06.784 00:07:07.757 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:07.757 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:07.757 Nvme0n1 : 2.00 25279.50 98.75 0.00 0.00 0.00 0.00 0.00 00:07:07.757 [2024-10-17T15:28:15.676Z] =================================================================================================================== 00:07:07.757 [2024-10-17T15:28:15.676Z] Total : 25279.50 98.75 0.00 0.00 0.00 0.00 0.00 00:07:07.757 00:07:07.757 true 00:07:07.757 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:07.757 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:07:08.017 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:07:08.017 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:07:08.017 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 4059986 00:07:08.956 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:08.956 Nvme0n1 : 3.00 25364.67 99.08 0.00 0.00 0.00 0.00 0.00 00:07:08.956 [2024-10-17T15:28:16.875Z] =================================================================================================================== 00:07:08.956 [2024-10-17T15:28:16.875Z] Total : 25364.67 99.08 0.00 0.00 0.00 0.00 0.00 00:07:08.956 00:07:09.895 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:09.895 Nvme0n1 : 4.00 25423.25 99.31 0.00 0.00 0.00 0.00 0.00 00:07:09.895 [2024-10-17T15:28:17.814Z] =================================================================================================================== 00:07:09.895 [2024-10-17T15:28:17.814Z] Total : 25423.25 99.31 0.00 0.00 0.00 0.00 0.00 00:07:09.895 00:07:10.837 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:10.837 Nvme0n1 : 5.00 25458.60 99.45 0.00 0.00 0.00 0.00 0.00 00:07:10.837 [2024-10-17T15:28:18.756Z] =================================================================================================================== 00:07:10.837 [2024-10-17T15:28:18.756Z] Total : 25458.60 99.45 0.00 0.00 0.00 0.00 0.00 00:07:10.837 00:07:11.777 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:11.777 Nvme0n1 : 6.00 25492.67 99.58 0.00 0.00 0.00 0.00 0.00 00:07:11.777 [2024-10-17T15:28:19.696Z] =================================================================================================================== 00:07:11.777 [2024-10-17T15:28:19.696Z] Total : 25492.67 99.58 0.00 0.00 0.00 0.00 0.00 00:07:11.777 00:07:12.716 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:12.716 Nvme0n1 : 7.00 25508.00 99.64 0.00 0.00 0.00 0.00 0.00 00:07:12.716 [2024-10-17T15:28:20.635Z] =================================================================================================================== 00:07:12.716 [2024-10-17T15:28:20.635Z] Total : 25508.00 99.64 0.00 0.00 0.00 0.00 0.00 00:07:12.716 00:07:13.656 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:13.656 Nvme0n1 : 8.00 25527.38 99.72 0.00 0.00 0.00 0.00 0.00 00:07:13.656 [2024-10-17T15:28:21.575Z] =================================================================================================================== 00:07:13.656 [2024-10-17T15:28:21.575Z] Total : 25527.38 99.72 0.00 0.00 0.00 0.00 0.00 00:07:13.656 00:07:15.038 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:15.038 Nvme0n1 : 9.00 25542.00 99.77 0.00 0.00 0.00 0.00 0.00 00:07:15.038 [2024-10-17T15:28:22.957Z] =================================================================================================================== 00:07:15.038 [2024-10-17T15:28:22.957Z] Total : 25542.00 99.77 0.00 0.00 0.00 0.00 0.00 00:07:15.038 00:07:15.979 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:15.979 Nvme0n1 : 10.00 25553.80 99.82 0.00 0.00 0.00 0.00 0.00 00:07:15.979 [2024-10-17T15:28:23.898Z] =================================================================================================================== 00:07:15.979 [2024-10-17T15:28:23.899Z] Total : 25553.80 99.82 0.00 0.00 0.00 0.00 0.00 00:07:15.980 00:07:15.980 00:07:15.980 Latency(us) 00:07:15.980 [2024-10-17T15:28:23.899Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:15.980 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:07:15.980 Nvme0n1 : 10.00 25558.80 99.84 0.00 0.00 5005.21 3085.65 14090.24 00:07:15.980 [2024-10-17T15:28:23.899Z] =================================================================================================================== 00:07:15.980 [2024-10-17T15:28:23.899Z] Total : 25558.80 99.84 0.00 0.00 5005.21 3085.65 14090.24 00:07:15.980 { 00:07:15.980 "results": [ 00:07:15.980 { 00:07:15.980 "job": "Nvme0n1", 00:07:15.980 "core_mask": "0x2", 00:07:15.980 "workload": "randwrite", 00:07:15.980 "status": "finished", 00:07:15.980 "queue_depth": 128, 00:07:15.980 "io_size": 4096, 00:07:15.980 "runtime": 10.003052, 00:07:15.980 "iops": 25558.799454406515, 00:07:15.980 "mibps": 99.83906036877545, 00:07:15.980 "io_failed": 0, 00:07:15.980 "io_timeout": 0, 00:07:15.980 "avg_latency_us": 5005.211864020506, 00:07:15.980 "min_latency_us": 3085.653333333333, 00:07:15.980 "max_latency_us": 14090.24 00:07:15.980 } 00:07:15.980 ], 00:07:15.980 "core_count": 1 00:07:15.980 } 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 4059913 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@950 -- # '[' -z 4059913 ']' 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # kill -0 4059913 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # uname 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4059913 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4059913' 00:07:15.980 killing process with pid 4059913 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@969 -- # kill 4059913 00:07:15.980 Received shutdown signal, test time was about 10.000000 seconds 00:07:15.980 00:07:15.980 Latency(us) 00:07:15.980 [2024-10-17T15:28:23.899Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:15.980 [2024-10-17T15:28:23.899Z] =================================================================================================================== 00:07:15.980 [2024-10-17T15:28:23.899Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@974 -- # wait 4059913 00:07:15.980 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:16.240 17:28:23 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:07:16.240 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:16.240 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 4056106 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 4056106 00:07:16.501 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 4056106 Killed "${NVMF_APP[@]}" "$@" 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@724 -- # xtrace_disable 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # nvmfpid=4062279 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # waitforlisten 4062279 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 4062279 ']' 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:16.501 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:16.502 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:16.502 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:16.502 17:28:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:07:16.502 [2024-10-17 17:28:24.402029] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:16.502 [2024-10-17 17:28:24.402083] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:16.762 [2024-10-17 17:28:24.484270] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:16.762 [2024-10-17 17:28:24.514301] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:16.762 [2024-10-17 17:28:24.514328] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:16.762 [2024-10-17 17:28:24.514334] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:16.762 [2024-10-17 17:28:24.514339] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:16.762 [2024-10-17 17:28:24.514343] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:16.762 [2024-10-17 17:28:24.514824] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:17.334 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:17.334 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:07:17.334 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:07:17.334 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@730 -- # xtrace_disable 00:07:17.334 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:07:17.334 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:17.334 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:07:17.595 [2024-10-17 17:28:25.376387] blobstore.c:4875:bs_recover: *NOTICE*: Performing recovery on blobstore 00:07:17.595 [2024-10-17 17:28:25.376506] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:07:17.595 [2024-10-17 17:28:25.376528] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev a4a89185-6e85-47b5-928c-08e39fe65533 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=a4a89185-6e85-47b5-928c-08e39fe65533 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:17.595 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:07:17.856 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b a4a89185-6e85-47b5-928c-08e39fe65533 -t 2000 00:07:17.856 [ 00:07:17.856 { 00:07:17.856 "name": "a4a89185-6e85-47b5-928c-08e39fe65533", 00:07:17.856 "aliases": [ 00:07:17.856 "lvs/lvol" 00:07:17.856 ], 00:07:17.856 "product_name": "Logical Volume", 00:07:17.856 "block_size": 4096, 00:07:17.856 "num_blocks": 38912, 00:07:17.856 "uuid": "a4a89185-6e85-47b5-928c-08e39fe65533", 00:07:17.856 "assigned_rate_limits": { 00:07:17.856 "rw_ios_per_sec": 0, 00:07:17.856 "rw_mbytes_per_sec": 0, 00:07:17.856 "r_mbytes_per_sec": 0, 00:07:17.856 "w_mbytes_per_sec": 0 00:07:17.856 }, 00:07:17.856 "claimed": false, 00:07:17.856 "zoned": false, 00:07:17.856 "supported_io_types": { 00:07:17.856 "read": true, 00:07:17.856 "write": true, 00:07:17.856 "unmap": true, 00:07:17.856 "flush": false, 00:07:17.856 "reset": true, 00:07:17.856 "nvme_admin": false, 00:07:17.856 "nvme_io": false, 00:07:17.856 "nvme_io_md": false, 00:07:17.856 "write_zeroes": true, 00:07:17.856 "zcopy": false, 00:07:17.856 "get_zone_info": false, 00:07:17.856 "zone_management": false, 00:07:17.856 "zone_append": false, 00:07:17.856 "compare": false, 00:07:17.856 "compare_and_write": false, 00:07:17.856 "abort": false, 00:07:17.856 "seek_hole": true, 00:07:17.856 "seek_data": true, 00:07:17.856 "copy": false, 00:07:17.856 "nvme_iov_md": false 00:07:17.856 }, 00:07:17.856 "driver_specific": { 00:07:17.856 "lvol": { 00:07:17.856 "lvol_store_uuid": "4d250600-c9ba-44c0-89d7-250e049e2a39", 00:07:17.856 "base_bdev": "aio_bdev", 00:07:17.856 "thin_provision": false, 00:07:17.856 "num_allocated_clusters": 38, 00:07:17.856 "snapshot": false, 00:07:17.856 "clone": false, 00:07:17.856 "esnap_clone": false 00:07:17.856 } 00:07:17.856 } 00:07:17.856 } 00:07:17.856 ] 00:07:17.856 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:07:17.856 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:17.856 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:07:18.117 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:07:18.117 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:18.117 17:28:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:07:18.378 [2024-10-17 17:28:26.212997] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@650 -- # local es=0 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:07:18.378 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:18.639 request: 00:07:18.639 { 00:07:18.639 "uuid": "4d250600-c9ba-44c0-89d7-250e049e2a39", 00:07:18.639 "method": "bdev_lvol_get_lvstores", 00:07:18.639 "req_id": 1 00:07:18.639 } 00:07:18.639 Got JSON-RPC error response 00:07:18.639 response: 00:07:18.639 { 00:07:18.639 "code": -19, 00:07:18.639 "message": "No such device" 00:07:18.639 } 00:07:18.639 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # es=1 00:07:18.639 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:18.639 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:18.639 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:18.639 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:07:18.899 aio_bdev 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev a4a89185-6e85-47b5-928c-08e39fe65533 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=a4a89185-6e85-47b5-928c-08e39fe65533 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:07:18.899 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b a4a89185-6e85-47b5-928c-08e39fe65533 -t 2000 00:07:19.160 [ 00:07:19.160 { 00:07:19.160 "name": "a4a89185-6e85-47b5-928c-08e39fe65533", 00:07:19.160 "aliases": [ 00:07:19.160 "lvs/lvol" 00:07:19.160 ], 00:07:19.160 "product_name": "Logical Volume", 00:07:19.160 "block_size": 4096, 00:07:19.160 "num_blocks": 38912, 00:07:19.160 "uuid": "a4a89185-6e85-47b5-928c-08e39fe65533", 00:07:19.160 "assigned_rate_limits": { 00:07:19.160 "rw_ios_per_sec": 0, 00:07:19.160 "rw_mbytes_per_sec": 0, 00:07:19.160 "r_mbytes_per_sec": 0, 00:07:19.160 "w_mbytes_per_sec": 0 00:07:19.160 }, 00:07:19.160 "claimed": false, 00:07:19.160 "zoned": false, 00:07:19.160 "supported_io_types": { 00:07:19.160 "read": true, 00:07:19.160 "write": true, 00:07:19.160 "unmap": true, 00:07:19.160 "flush": false, 00:07:19.160 "reset": true, 00:07:19.160 "nvme_admin": false, 00:07:19.160 "nvme_io": false, 00:07:19.160 "nvme_io_md": false, 00:07:19.160 "write_zeroes": true, 00:07:19.160 "zcopy": false, 00:07:19.160 "get_zone_info": false, 00:07:19.160 "zone_management": false, 00:07:19.160 "zone_append": false, 00:07:19.160 "compare": false, 00:07:19.160 "compare_and_write": false, 00:07:19.160 "abort": false, 00:07:19.160 "seek_hole": true, 00:07:19.160 "seek_data": true, 00:07:19.160 "copy": false, 00:07:19.160 "nvme_iov_md": false 00:07:19.160 }, 00:07:19.160 "driver_specific": { 00:07:19.160 "lvol": { 00:07:19.160 "lvol_store_uuid": "4d250600-c9ba-44c0-89d7-250e049e2a39", 00:07:19.160 "base_bdev": "aio_bdev", 00:07:19.160 "thin_provision": false, 00:07:19.160 "num_allocated_clusters": 38, 00:07:19.160 "snapshot": false, 00:07:19.160 "clone": false, 00:07:19.160 "esnap_clone": false 00:07:19.160 } 00:07:19.160 } 00:07:19.160 } 00:07:19.160 ] 00:07:19.160 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:07:19.160 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:19.160 17:28:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:07:19.421 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:07:19.421 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:19.421 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:07:19.421 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:07:19.421 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete a4a89185-6e85-47b5-928c-08e39fe65533 00:07:19.681 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 4d250600-c9ba-44c0-89d7-250e049e2a39 00:07:19.940 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:07:19.940 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:07:19.940 00:07:19.940 real 0m16.791s 00:07:19.940 user 0m44.253s 00:07:19.940 sys 0m3.064s 00:07:19.940 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:19.940 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:07:19.940 ************************************ 00:07:19.940 END TEST lvs_grow_dirty 00:07:19.940 ************************************ 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@808 -- # type=--id 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@809 -- # id=0 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # for n in $shm_files 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:07:20.199 nvmf_trace.0 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@823 -- # return 0 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@514 -- # nvmfcleanup 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:07:20.199 rmmod nvme_tcp 00:07:20.199 rmmod nvme_fabrics 00:07:20.199 rmmod nvme_keyring 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@515 -- # '[' -n 4062279 ']' 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@516 -- # killprocess 4062279 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@950 -- # '[' -z 4062279 ']' 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # kill -0 4062279 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # uname 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:20.199 17:28:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4062279 00:07:20.199 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:20.199 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:20.199 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4062279' 00:07:20.199 killing process with pid 4062279 00:07:20.199 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@969 -- # kill 4062279 00:07:20.199 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@974 -- # wait 4062279 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-save 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-restore 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:07:20.459 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:07:20.460 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:20.460 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:20.460 17:28:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:22.371 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:07:22.371 00:07:22.371 real 0m44.059s 00:07:22.371 user 1m6.248s 00:07:22.371 sys 0m10.562s 00:07:22.371 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:22.371 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:07:22.371 ************************************ 00:07:22.371 END TEST nvmf_lvs_grow 00:07:22.371 ************************************ 00:07:22.371 17:28:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:07:22.371 17:28:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:22.371 17:28:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:22.371 17:28:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:07:22.632 ************************************ 00:07:22.632 START TEST nvmf_bdev_io_wait 00:07:22.632 ************************************ 00:07:22.632 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:07:22.632 * Looking for test storage... 00:07:22.632 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1691 -- # lcov --version 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:07:22.633 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.633 --rc genhtml_branch_coverage=1 00:07:22.633 --rc genhtml_function_coverage=1 00:07:22.633 --rc genhtml_legend=1 00:07:22.633 --rc geninfo_all_blocks=1 00:07:22.633 --rc geninfo_unexecuted_blocks=1 00:07:22.633 00:07:22.633 ' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:07:22.633 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.633 --rc genhtml_branch_coverage=1 00:07:22.633 --rc genhtml_function_coverage=1 00:07:22.633 --rc genhtml_legend=1 00:07:22.633 --rc geninfo_all_blocks=1 00:07:22.633 --rc geninfo_unexecuted_blocks=1 00:07:22.633 00:07:22.633 ' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:07:22.633 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.633 --rc genhtml_branch_coverage=1 00:07:22.633 --rc genhtml_function_coverage=1 00:07:22.633 --rc genhtml_legend=1 00:07:22.633 --rc geninfo_all_blocks=1 00:07:22.633 --rc geninfo_unexecuted_blocks=1 00:07:22.633 00:07:22.633 ' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:07:22.633 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:22.633 --rc genhtml_branch_coverage=1 00:07:22.633 --rc genhtml_function_coverage=1 00:07:22.633 --rc genhtml_legend=1 00:07:22.633 --rc geninfo_all_blocks=1 00:07:22.633 --rc geninfo_unexecuted_blocks=1 00:07:22.633 00:07:22.633 ' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:22.633 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:07:22.895 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # prepare_net_devs 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@436 -- # local -g is_hw=no 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # remove_spdk_ns 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:07:22.895 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:07:31.035 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:07:31.036 Found 0000:31:00.0 (0x8086 - 0x159b) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:07:31.036 Found 0000:31:00.1 (0x8086 - 0x159b) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:07:31.036 Found net devices under 0000:31:00.0: cvl_0_0 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:07:31.036 Found net devices under 0000:31:00.1: cvl_0_1 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # is_hw=yes 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:07:31.036 17:28:37 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:07:31.036 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:31.036 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.640 ms 00:07:31.036 00:07:31.036 --- 10.0.0.2 ping statistics --- 00:07:31.036 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:31.036 rtt min/avg/max/mdev = 0.640/0.640/0.640/0.000 ms 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:31.036 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:31.036 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.317 ms 00:07:31.036 00:07:31.036 --- 10.0.0.1 ping statistics --- 00:07:31.036 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:31.036 rtt min/avg/max/mdev = 0.317/0.317/0.317/0.000 ms 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@448 -- # return 0 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@724 -- # xtrace_disable 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # nvmfpid=4067387 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # waitforlisten 4067387 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@831 -- # '[' -z 4067387 ']' 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:31.036 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:31.036 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.036 [2024-10-17 17:28:38.274637] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:31.036 [2024-10-17 17:28:38.274707] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:31.036 [2024-10-17 17:28:38.363907] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:31.036 [2024-10-17 17:28:38.419149] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:31.036 [2024-10-17 17:28:38.419198] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:31.036 [2024-10-17 17:28:38.419207] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:31.036 [2024-10-17 17:28:38.419214] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:31.036 [2024-10-17 17:28:38.419221] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:31.036 [2024-10-17 17:28:38.421627] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:31.036 [2024-10-17 17:28:38.421761] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:31.036 [2024-10-17 17:28:38.421934] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:07:31.036 [2024-10-17 17:28:38.421936] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # return 0 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@730 -- # xtrace_disable 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.297 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.559 [2024-10-17 17:28:39.225666] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.559 Malloc0 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:31.559 [2024-10-17 17:28:39.291334] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=4067494 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=4067497 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:07:31.559 { 00:07:31.559 "params": { 00:07:31.559 "name": "Nvme$subsystem", 00:07:31.559 "trtype": "$TEST_TRANSPORT", 00:07:31.559 "traddr": "$NVMF_FIRST_TARGET_IP", 00:07:31.559 "adrfam": "ipv4", 00:07:31.559 "trsvcid": "$NVMF_PORT", 00:07:31.559 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:07:31.559 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:07:31.559 "hdgst": ${hdgst:-false}, 00:07:31.559 "ddgst": ${ddgst:-false} 00:07:31.559 }, 00:07:31.559 "method": "bdev_nvme_attach_controller" 00:07:31.559 } 00:07:31.559 EOF 00:07:31.559 )") 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=4067500 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=4067504 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:07:31.559 { 00:07:31.559 "params": { 00:07:31.559 "name": "Nvme$subsystem", 00:07:31.559 "trtype": "$TEST_TRANSPORT", 00:07:31.559 "traddr": "$NVMF_FIRST_TARGET_IP", 00:07:31.559 "adrfam": "ipv4", 00:07:31.559 "trsvcid": "$NVMF_PORT", 00:07:31.559 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:07:31.559 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:07:31.559 "hdgst": ${hdgst:-false}, 00:07:31.559 "ddgst": ${ddgst:-false} 00:07:31.559 }, 00:07:31.559 "method": "bdev_nvme_attach_controller" 00:07:31.559 } 00:07:31.559 EOF 00:07:31.559 )") 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:07:31.559 { 00:07:31.559 "params": { 00:07:31.559 "name": "Nvme$subsystem", 00:07:31.559 "trtype": "$TEST_TRANSPORT", 00:07:31.559 "traddr": "$NVMF_FIRST_TARGET_IP", 00:07:31.559 "adrfam": "ipv4", 00:07:31.559 "trsvcid": "$NVMF_PORT", 00:07:31.559 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:07:31.559 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:07:31.559 "hdgst": ${hdgst:-false}, 00:07:31.559 "ddgst": ${ddgst:-false} 00:07:31.559 }, 00:07:31.559 "method": "bdev_nvme_attach_controller" 00:07:31.559 } 00:07:31.559 EOF 00:07:31.559 )") 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:07:31.559 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:07:31.560 { 00:07:31.560 "params": { 00:07:31.560 "name": "Nvme$subsystem", 00:07:31.560 "trtype": "$TEST_TRANSPORT", 00:07:31.560 "traddr": "$NVMF_FIRST_TARGET_IP", 00:07:31.560 "adrfam": "ipv4", 00:07:31.560 "trsvcid": "$NVMF_PORT", 00:07:31.560 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:07:31.560 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:07:31.560 "hdgst": ${hdgst:-false}, 00:07:31.560 "ddgst": ${ddgst:-false} 00:07:31.560 }, 00:07:31.560 "method": "bdev_nvme_attach_controller" 00:07:31.560 } 00:07:31.560 EOF 00:07:31.560 )") 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 4067494 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:07:31.560 "params": { 00:07:31.560 "name": "Nvme1", 00:07:31.560 "trtype": "tcp", 00:07:31.560 "traddr": "10.0.0.2", 00:07:31.560 "adrfam": "ipv4", 00:07:31.560 "trsvcid": "4420", 00:07:31.560 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:07:31.560 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:07:31.560 "hdgst": false, 00:07:31.560 "ddgst": false 00:07:31.560 }, 00:07:31.560 "method": "bdev_nvme_attach_controller" 00:07:31.560 }' 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:07:31.560 "params": { 00:07:31.560 "name": "Nvme1", 00:07:31.560 "trtype": "tcp", 00:07:31.560 "traddr": "10.0.0.2", 00:07:31.560 "adrfam": "ipv4", 00:07:31.560 "trsvcid": "4420", 00:07:31.560 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:07:31.560 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:07:31.560 "hdgst": false, 00:07:31.560 "ddgst": false 00:07:31.560 }, 00:07:31.560 "method": "bdev_nvme_attach_controller" 00:07:31.560 }' 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:07:31.560 "params": { 00:07:31.560 "name": "Nvme1", 00:07:31.560 "trtype": "tcp", 00:07:31.560 "traddr": "10.0.0.2", 00:07:31.560 "adrfam": "ipv4", 00:07:31.560 "trsvcid": "4420", 00:07:31.560 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:07:31.560 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:07:31.560 "hdgst": false, 00:07:31.560 "ddgst": false 00:07:31.560 }, 00:07:31.560 "method": "bdev_nvme_attach_controller" 00:07:31.560 }' 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:07:31.560 17:28:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:07:31.560 "params": { 00:07:31.560 "name": "Nvme1", 00:07:31.560 "trtype": "tcp", 00:07:31.560 "traddr": "10.0.0.2", 00:07:31.560 "adrfam": "ipv4", 00:07:31.560 "trsvcid": "4420", 00:07:31.560 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:07:31.560 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:07:31.560 "hdgst": false, 00:07:31.560 "ddgst": false 00:07:31.560 }, 00:07:31.560 "method": "bdev_nvme_attach_controller" 00:07:31.560 }' 00:07:31.560 [2024-10-17 17:28:39.349829] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:31.560 [2024-10-17 17:28:39.349905] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:07:31.560 [2024-10-17 17:28:39.352566] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:31.560 [2024-10-17 17:28:39.352627] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:07:31.560 [2024-10-17 17:28:39.353070] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:31.560 [2024-10-17 17:28:39.353137] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:07:31.560 [2024-10-17 17:28:39.360372] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:31.560 [2024-10-17 17:28:39.360456] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:07:31.821 [2024-10-17 17:28:39.534867] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:31.821 [2024-10-17 17:28:39.570669] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:07:31.821 [2024-10-17 17:28:39.597306] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:31.821 [2024-10-17 17:28:39.632883] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:07:31.821 [2024-10-17 17:28:39.659006] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:31.821 [2024-10-17 17:28:39.699045] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:07:32.082 [2024-10-17 17:28:39.754248] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:32.082 [2024-10-17 17:28:39.797805] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:07:32.082 Running I/O for 1 seconds... 00:07:32.342 Running I/O for 1 seconds... 00:07:32.342 Running I/O for 1 seconds... 00:07:32.342 Running I/O for 1 seconds... 00:07:33.283 10809.00 IOPS, 42.22 MiB/s 00:07:33.283 Latency(us) 00:07:33.283 [2024-10-17T15:28:41.202Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:33.283 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:07:33.283 Nvme1n1 : 1.01 10855.47 42.40 0.00 0.00 11745.94 6335.15 17039.36 00:07:33.283 [2024-10-17T15:28:41.202Z] =================================================================================================================== 00:07:33.283 [2024-10-17T15:28:41.202Z] Total : 10855.47 42.40 0.00 0.00 11745.94 6335.15 17039.36 00:07:33.283 186352.00 IOPS, 727.94 MiB/s [2024-10-17T15:28:41.202Z] 10663.00 IOPS, 41.65 MiB/s 00:07:33.283 Latency(us) 00:07:33.283 [2024-10-17T15:28:41.202Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:33.283 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:07:33.283 Nvme1n1 : 1.00 185983.38 726.50 0.00 0.00 684.69 305.49 1979.73 00:07:33.283 [2024-10-17T15:28:41.202Z] =================================================================================================================== 00:07:33.283 [2024-10-17T15:28:41.202Z] Total : 185983.38 726.50 0.00 0.00 684.69 305.49 1979.73 00:07:33.283 00:07:33.283 Latency(us) 00:07:33.283 [2024-10-17T15:28:41.202Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:33.283 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:07:33.283 Nvme1n1 : 1.01 10731.00 41.92 0.00 0.00 11884.99 5297.49 20862.29 00:07:33.283 [2024-10-17T15:28:41.202Z] =================================================================================================================== 00:07:33.283 [2024-10-17T15:28:41.202Z] Total : 10731.00 41.92 0.00 0.00 11884.99 5297.49 20862.29 00:07:33.283 9287.00 IOPS, 36.28 MiB/s 00:07:33.283 Latency(us) 00:07:33.283 [2024-10-17T15:28:41.202Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:33.283 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:07:33.283 Nvme1n1 : 1.01 9370.46 36.60 0.00 0.00 13614.05 4860.59 24357.55 00:07:33.283 [2024-10-17T15:28:41.202Z] =================================================================================================================== 00:07:33.283 [2024-10-17T15:28:41.202Z] Total : 9370.46 36.60 0.00 0.00 13614.05 4860.59 24357.55 00:07:33.543 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 4067497 00:07:33.543 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 4067500 00:07:33.543 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 4067504 00:07:33.543 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:07:33.543 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@514 -- # nvmfcleanup 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:07:33.544 rmmod nvme_tcp 00:07:33.544 rmmod nvme_fabrics 00:07:33.544 rmmod nvme_keyring 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@515 -- # '[' -n 4067387 ']' 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # killprocess 4067387 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@950 -- # '[' -z 4067387 ']' 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # kill -0 4067387 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # uname 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4067387 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4067387' 00:07:33.544 killing process with pid 4067387 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@969 -- # kill 4067387 00:07:33.544 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@974 -- # wait 4067387 00:07:33.804 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-save 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-restore 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:33.805 17:28:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:35.715 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:07:35.715 00:07:35.715 real 0m13.247s 00:07:35.715 user 0m19.998s 00:07:35.715 sys 0m7.498s 00:07:35.715 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:35.715 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:07:35.715 ************************************ 00:07:35.715 END TEST nvmf_bdev_io_wait 00:07:35.715 ************************************ 00:07:35.715 17:28:43 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:07:35.715 17:28:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:35.715 17:28:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:35.715 17:28:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:07:35.976 ************************************ 00:07:35.976 START TEST nvmf_queue_depth 00:07:35.976 ************************************ 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:07:35.976 * Looking for test storage... 00:07:35.976 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1691 -- # lcov --version 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:07:35.976 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.976 --rc genhtml_branch_coverage=1 00:07:35.976 --rc genhtml_function_coverage=1 00:07:35.976 --rc genhtml_legend=1 00:07:35.976 --rc geninfo_all_blocks=1 00:07:35.976 --rc geninfo_unexecuted_blocks=1 00:07:35.976 00:07:35.976 ' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:07:35.976 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.976 --rc genhtml_branch_coverage=1 00:07:35.976 --rc genhtml_function_coverage=1 00:07:35.976 --rc genhtml_legend=1 00:07:35.976 --rc geninfo_all_blocks=1 00:07:35.976 --rc geninfo_unexecuted_blocks=1 00:07:35.976 00:07:35.976 ' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:07:35.976 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.976 --rc genhtml_branch_coverage=1 00:07:35.976 --rc genhtml_function_coverage=1 00:07:35.976 --rc genhtml_legend=1 00:07:35.976 --rc geninfo_all_blocks=1 00:07:35.976 --rc geninfo_unexecuted_blocks=1 00:07:35.976 00:07:35.976 ' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:07:35.976 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.976 --rc genhtml_branch_coverage=1 00:07:35.976 --rc genhtml_function_coverage=1 00:07:35.976 --rc genhtml_legend=1 00:07:35.976 --rc geninfo_all_blocks=1 00:07:35.976 --rc geninfo_unexecuted_blocks=1 00:07:35.976 00:07:35.976 ' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:35.976 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:07:35.977 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@474 -- # prepare_net_devs 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@436 -- # local -g is_hw=no 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@438 -- # remove_spdk_ns 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:35.977 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:36.238 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:07:36.238 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:07:36.238 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:07:36.238 17:28:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:07:44.382 Found 0000:31:00.0 (0x8086 - 0x159b) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:07:44.382 Found 0000:31:00.1 (0x8086 - 0x159b) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:07:44.382 Found net devices under 0000:31:00.0: cvl_0_0 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:07:44.382 Found net devices under 0000:31:00.1: cvl_0_1 00:07:44.382 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # is_hw=yes 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:07:44.383 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:44.383 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.688 ms 00:07:44.383 00:07:44.383 --- 10.0.0.2 ping statistics --- 00:07:44.383 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:44.383 rtt min/avg/max/mdev = 0.688/0.688/0.688/0.000 ms 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:44.383 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:44.383 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.283 ms 00:07:44.383 00:07:44.383 --- 10.0.0.1 ping statistics --- 00:07:44.383 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:44.383 rtt min/avg/max/mdev = 0.283/0.283/0.283/0.000 ms 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@448 -- # return 0 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@724 -- # xtrace_disable 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@507 -- # nvmfpid=4072220 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@508 -- # waitforlisten 4072220 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 4072220 ']' 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:44.383 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:44.383 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.383 [2024-10-17 17:28:51.614822] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:44.383 [2024-10-17 17:28:51.614892] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:44.383 [2024-10-17 17:28:51.707220] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:44.383 [2024-10-17 17:28:51.759388] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:44.383 [2024-10-17 17:28:51.759446] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:44.383 [2024-10-17 17:28:51.759455] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:44.383 [2024-10-17 17:28:51.759462] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:44.383 [2024-10-17 17:28:51.759469] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:44.383 [2024-10-17 17:28:51.760295] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@730 -- # xtrace_disable 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.648 [2024-10-17 17:28:52.499894] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.648 Malloc0 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.648 [2024-10-17 17:28:52.561245] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:44.648 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=4072508 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 4072508 /var/tmp/bdevperf.sock 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 4072508 ']' 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:07:44.909 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:44.909 17:28:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:44.909 [2024-10-17 17:28:52.627655] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:07:44.909 [2024-10-17 17:28:52.627733] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4072508 ] 00:07:44.909 [2024-10-17 17:28:52.711727] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:44.909 [2024-10-17 17:28:52.765338] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:45.850 17:28:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:45.850 17:28:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:07:45.850 17:28:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:07:45.850 17:28:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.850 17:28:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:45.850 NVMe0n1 00:07:45.850 17:28:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.850 17:28:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:07:45.850 Running I/O for 10 seconds... 00:07:48.171 9095.00 IOPS, 35.53 MiB/s [2024-10-17T15:28:57.030Z] 10240.50 IOPS, 40.00 MiB/s [2024-10-17T15:28:57.967Z] 10721.00 IOPS, 41.88 MiB/s [2024-10-17T15:28:58.907Z] 11228.50 IOPS, 43.86 MiB/s [2024-10-17T15:28:59.842Z] 11637.80 IOPS, 45.46 MiB/s [2024-10-17T15:29:01.220Z] 11905.83 IOPS, 46.51 MiB/s [2024-10-17T15:29:02.160Z] 12066.86 IOPS, 47.14 MiB/s [2024-10-17T15:29:03.101Z] 12238.50 IOPS, 47.81 MiB/s [2024-10-17T15:29:04.039Z] 12334.44 IOPS, 48.18 MiB/s [2024-10-17T15:29:04.039Z] 12462.50 IOPS, 48.68 MiB/s 00:07:56.120 Latency(us) 00:07:56.120 [2024-10-17T15:29:04.039Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:56.120 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:07:56.120 Verification LBA range: start 0x0 length 0x4000 00:07:56.120 NVMe0n1 : 10.05 12482.82 48.76 0.00 0.00 81719.69 18022.40 77332.48 00:07:56.120 [2024-10-17T15:29:04.039Z] =================================================================================================================== 00:07:56.120 [2024-10-17T15:29:04.039Z] Total : 12482.82 48.76 0.00 0.00 81719.69 18022.40 77332.48 00:07:56.120 { 00:07:56.120 "results": [ 00:07:56.120 { 00:07:56.120 "job": "NVMe0n1", 00:07:56.120 "core_mask": "0x1", 00:07:56.120 "workload": "verify", 00:07:56.120 "status": "finished", 00:07:56.120 "verify_range": { 00:07:56.120 "start": 0, 00:07:56.120 "length": 16384 00:07:56.120 }, 00:07:56.120 "queue_depth": 1024, 00:07:56.120 "io_size": 4096, 00:07:56.120 "runtime": 10.0539, 00:07:56.120 "iops": 12482.817613065576, 00:07:56.120 "mibps": 48.761006301037405, 00:07:56.120 "io_failed": 0, 00:07:56.120 "io_timeout": 0, 00:07:56.120 "avg_latency_us": 81719.68989824782, 00:07:56.120 "min_latency_us": 18022.4, 00:07:56.120 "max_latency_us": 77332.48 00:07:56.120 } 00:07:56.120 ], 00:07:56.120 "core_count": 1 00:07:56.120 } 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 4072508 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 4072508 ']' 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 4072508 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4072508 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4072508' 00:07:56.120 killing process with pid 4072508 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 4072508 00:07:56.120 Received shutdown signal, test time was about 10.000000 seconds 00:07:56.120 00:07:56.120 Latency(us) 00:07:56.120 [2024-10-17T15:29:04.039Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:07:56.120 [2024-10-17T15:29:04.039Z] =================================================================================================================== 00:07:56.120 [2024-10-17T15:29:04.039Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:07:56.120 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 4072508 00:07:56.120 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:07:56.120 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:07:56.380 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@514 -- # nvmfcleanup 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:07:56.381 rmmod nvme_tcp 00:07:56.381 rmmod nvme_fabrics 00:07:56.381 rmmod nvme_keyring 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@515 -- # '[' -n 4072220 ']' 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@516 -- # killprocess 4072220 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 4072220 ']' 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 4072220 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4072220 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4072220' 00:07:56.381 killing process with pid 4072220 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 4072220 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 4072220 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-save 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:07:56.381 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-restore 00:07:56.641 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:07:56.641 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:07:56.641 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:56.641 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:56.641 17:29:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:07:58.552 00:07:58.552 real 0m22.723s 00:07:58.552 user 0m26.103s 00:07:58.552 sys 0m7.034s 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:07:58.552 ************************************ 00:07:58.552 END TEST nvmf_queue_depth 00:07:58.552 ************************************ 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:07:58.552 ************************************ 00:07:58.552 START TEST nvmf_target_multipath 00:07:58.552 ************************************ 00:07:58.552 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:07:58.813 * Looking for test storage... 00:07:58.813 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1691 -- # lcov --version 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:07:58.813 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:07:58.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:58.814 --rc genhtml_branch_coverage=1 00:07:58.814 --rc genhtml_function_coverage=1 00:07:58.814 --rc genhtml_legend=1 00:07:58.814 --rc geninfo_all_blocks=1 00:07:58.814 --rc geninfo_unexecuted_blocks=1 00:07:58.814 00:07:58.814 ' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:07:58.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:58.814 --rc genhtml_branch_coverage=1 00:07:58.814 --rc genhtml_function_coverage=1 00:07:58.814 --rc genhtml_legend=1 00:07:58.814 --rc geninfo_all_blocks=1 00:07:58.814 --rc geninfo_unexecuted_blocks=1 00:07:58.814 00:07:58.814 ' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:07:58.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:58.814 --rc genhtml_branch_coverage=1 00:07:58.814 --rc genhtml_function_coverage=1 00:07:58.814 --rc genhtml_legend=1 00:07:58.814 --rc geninfo_all_blocks=1 00:07:58.814 --rc geninfo_unexecuted_blocks=1 00:07:58.814 00:07:58.814 ' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:07:58.814 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:58.814 --rc genhtml_branch_coverage=1 00:07:58.814 --rc genhtml_function_coverage=1 00:07:58.814 --rc genhtml_legend=1 00:07:58.814 --rc geninfo_all_blocks=1 00:07:58.814 --rc geninfo_unexecuted_blocks=1 00:07:58.814 00:07:58.814 ' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:07:58.814 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@474 -- # prepare_net_devs 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@436 -- # local -g is_hw=no 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@438 -- # remove_spdk_ns 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:07:58.814 17:29:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:08:07.125 Found 0000:31:00.0 (0x8086 - 0x159b) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:08:07.125 Found 0000:31:00.1 (0x8086 - 0x159b) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:08:07.125 Found net devices under 0000:31:00.0: cvl_0_0 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:08:07.125 Found net devices under 0000:31:00.1: cvl_0_1 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # is_hw=yes 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:08:07.125 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:08:07.126 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:08:07.126 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:07.126 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.594 ms 00:08:07.126 00:08:07.126 --- 10.0.0.2 ping statistics --- 00:08:07.126 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:07.126 rtt min/avg/max/mdev = 0.594/0.594/0.594/0.000 ms 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:07.126 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:07.126 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.296 ms 00:08:07.126 00:08:07.126 --- 10.0.0.1 ping statistics --- 00:08:07.126 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:07.126 rtt min/avg/max/mdev = 0.296/0.296/0.296/0.000 ms 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@448 -- # return 0 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:08:07.126 only one NIC for nvmf test 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:07.126 rmmod nvme_tcp 00:08:07.126 rmmod nvme_fabrics 00:08:07.126 rmmod nvme_keyring 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:07.126 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:09.043 00:08:09.043 real 0m10.087s 00:08:09.043 user 0m2.157s 00:08:09.043 sys 0m5.857s 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:08:09.043 ************************************ 00:08:09.043 END TEST nvmf_target_multipath 00:08:09.043 ************************************ 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:08:09.043 ************************************ 00:08:09.043 START TEST nvmf_zcopy 00:08:09.043 ************************************ 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:08:09.043 * Looking for test storage... 00:08:09.043 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1691 -- # lcov --version 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:08:09.043 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.043 --rc genhtml_branch_coverage=1 00:08:09.043 --rc genhtml_function_coverage=1 00:08:09.043 --rc genhtml_legend=1 00:08:09.043 --rc geninfo_all_blocks=1 00:08:09.043 --rc geninfo_unexecuted_blocks=1 00:08:09.043 00:08:09.043 ' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:08:09.043 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.043 --rc genhtml_branch_coverage=1 00:08:09.043 --rc genhtml_function_coverage=1 00:08:09.043 --rc genhtml_legend=1 00:08:09.043 --rc geninfo_all_blocks=1 00:08:09.043 --rc geninfo_unexecuted_blocks=1 00:08:09.043 00:08:09.043 ' 00:08:09.043 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:08:09.043 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.043 --rc genhtml_branch_coverage=1 00:08:09.043 --rc genhtml_function_coverage=1 00:08:09.043 --rc genhtml_legend=1 00:08:09.043 --rc geninfo_all_blocks=1 00:08:09.043 --rc geninfo_unexecuted_blocks=1 00:08:09.043 00:08:09.044 ' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:08:09.044 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.044 --rc genhtml_branch_coverage=1 00:08:09.044 --rc genhtml_function_coverage=1 00:08:09.044 --rc genhtml_legend=1 00:08:09.044 --rc geninfo_all_blocks=1 00:08:09.044 --rc geninfo_unexecuted_blocks=1 00:08:09.044 00:08:09.044 ' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:09.044 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@474 -- # prepare_net_devs 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@436 -- # local -g is_hw=no 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@438 -- # remove_spdk_ns 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:08:09.044 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:08:17.187 Found 0000:31:00.0 (0x8086 - 0x159b) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:08:17.187 Found 0000:31:00.1 (0x8086 - 0x159b) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:08:17.187 Found net devices under 0000:31:00.0: cvl_0_0 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:17.187 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:08:17.188 Found net devices under 0000:31:00.1: cvl_0_1 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # is_hw=yes 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:08:17.188 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:17.188 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.607 ms 00:08:17.188 00:08:17.188 --- 10.0.0.2 ping statistics --- 00:08:17.188 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:17.188 rtt min/avg/max/mdev = 0.607/0.607/0.607/0.000 ms 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:17.188 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:17.188 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.325 ms 00:08:17.188 00:08:17.188 --- 10.0.0.1 ping statistics --- 00:08:17.188 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:17.188 rtt min/avg/max/mdev = 0.325/0.325/0.325/0.000 ms 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@448 -- # return 0 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@507 -- # nvmfpid=4083291 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@508 -- # waitforlisten 4083291 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@831 -- # '[' -z 4083291 ']' 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:17.188 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:17.188 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.188 [2024-10-17 17:29:24.598152] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:08:17.188 [2024-10-17 17:29:24.598221] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:17.188 [2024-10-17 17:29:24.687262] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.188 [2024-10-17 17:29:24.738245] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:17.188 [2024-10-17 17:29:24.738297] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:17.188 [2024-10-17 17:29:24.738307] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:17.188 [2024-10-17 17:29:24.738314] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:17.188 [2024-10-17 17:29:24.738321] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:17.188 [2024-10-17 17:29:24.739121] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:17.760 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:17.760 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@864 -- # return 0 00:08:17.760 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:08:17.760 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:17.760 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.760 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:17.760 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.761 [2024-10-17 17:29:25.481113] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.761 [2024-10-17 17:29:25.505400] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.761 malloc0 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:08:17.761 { 00:08:17.761 "params": { 00:08:17.761 "name": "Nvme$subsystem", 00:08:17.761 "trtype": "$TEST_TRANSPORT", 00:08:17.761 "traddr": "$NVMF_FIRST_TARGET_IP", 00:08:17.761 "adrfam": "ipv4", 00:08:17.761 "trsvcid": "$NVMF_PORT", 00:08:17.761 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:08:17.761 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:08:17.761 "hdgst": ${hdgst:-false}, 00:08:17.761 "ddgst": ${ddgst:-false} 00:08:17.761 }, 00:08:17.761 "method": "bdev_nvme_attach_controller" 00:08:17.761 } 00:08:17.761 EOF 00:08:17.761 )") 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:08:17.761 17:29:25 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:08:17.761 "params": { 00:08:17.761 "name": "Nvme1", 00:08:17.761 "trtype": "tcp", 00:08:17.761 "traddr": "10.0.0.2", 00:08:17.761 "adrfam": "ipv4", 00:08:17.761 "trsvcid": "4420", 00:08:17.761 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:08:17.761 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:08:17.761 "hdgst": false, 00:08:17.761 "ddgst": false 00:08:17.761 }, 00:08:17.761 "method": "bdev_nvme_attach_controller" 00:08:17.761 }' 00:08:17.761 [2024-10-17 17:29:25.615356] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:08:17.761 [2024-10-17 17:29:25.615418] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4083624 ] 00:08:18.021 [2024-10-17 17:29:25.698033] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:18.021 [2024-10-17 17:29:25.750876] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.281 Running I/O for 10 seconds... 00:08:20.163 6768.00 IOPS, 52.88 MiB/s [2024-10-17T15:29:29.023Z] 8201.00 IOPS, 64.07 MiB/s [2024-10-17T15:29:30.406Z] 8683.67 IOPS, 67.84 MiB/s [2024-10-17T15:29:30.977Z] 8928.00 IOPS, 69.75 MiB/s [2024-10-17T15:29:32.360Z] 9055.00 IOPS, 70.74 MiB/s [2024-10-17T15:29:33.301Z] 9159.67 IOPS, 71.56 MiB/s [2024-10-17T15:29:34.242Z] 9242.00 IOPS, 72.20 MiB/s [2024-10-17T15:29:35.182Z] 9307.00 IOPS, 72.71 MiB/s [2024-10-17T15:29:36.124Z] 9354.11 IOPS, 73.08 MiB/s [2024-10-17T15:29:36.124Z] 9392.90 IOPS, 73.38 MiB/s 00:08:28.205 Latency(us) 00:08:28.205 [2024-10-17T15:29:36.124Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:28.205 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:08:28.205 Verification LBA range: start 0x0 length 0x1000 00:08:28.205 Nvme1n1 : 10.01 9394.51 73.39 0.00 0.00 13577.24 1146.88 28617.39 00:08:28.205 [2024-10-17T15:29:36.124Z] =================================================================================================================== 00:08:28.205 [2024-10-17T15:29:36.124Z] Total : 9394.51 73.39 0.00 0.00 13577.24 1146.88 28617.39 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=4085640 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:08:28.205 { 00:08:28.205 "params": { 00:08:28.205 "name": "Nvme$subsystem", 00:08:28.205 "trtype": "$TEST_TRANSPORT", 00:08:28.205 "traddr": "$NVMF_FIRST_TARGET_IP", 00:08:28.205 "adrfam": "ipv4", 00:08:28.205 "trsvcid": "$NVMF_PORT", 00:08:28.205 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:08:28.205 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:08:28.205 "hdgst": ${hdgst:-false}, 00:08:28.205 "ddgst": ${ddgst:-false} 00:08:28.205 }, 00:08:28.205 "method": "bdev_nvme_attach_controller" 00:08:28.205 } 00:08:28.205 EOF 00:08:28.205 )") 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:08:28.205 [2024-10-17 17:29:36.095840] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.205 [2024-10-17 17:29:36.095873] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:08:28.205 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:08:28.205 "params": { 00:08:28.205 "name": "Nvme1", 00:08:28.205 "trtype": "tcp", 00:08:28.205 "traddr": "10.0.0.2", 00:08:28.205 "adrfam": "ipv4", 00:08:28.205 "trsvcid": "4420", 00:08:28.205 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:08:28.205 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:08:28.205 "hdgst": false, 00:08:28.205 "ddgst": false 00:08:28.205 }, 00:08:28.205 "method": "bdev_nvme_attach_controller" 00:08:28.205 }' 00:08:28.206 [2024-10-17 17:29:36.107831] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.206 [2024-10-17 17:29:36.107843] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.206 [2024-10-17 17:29:36.119856] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.206 [2024-10-17 17:29:36.119864] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.466 [2024-10-17 17:29:36.131888] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.466 [2024-10-17 17:29:36.131896] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.466 [2024-10-17 17:29:36.143917] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.466 [2024-10-17 17:29:36.143925] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.466 [2024-10-17 17:29:36.146778] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:08:28.466 [2024-10-17 17:29:36.146841] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4085640 ] 00:08:28.466 [2024-10-17 17:29:36.155948] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.466 [2024-10-17 17:29:36.155956] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.466 [2024-10-17 17:29:36.167981] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.466 [2024-10-17 17:29:36.167989] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.180010] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.180018] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.192041] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.192048] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.204072] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.204080] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.216104] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.216112] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.226169] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:28.467 [2024-10-17 17:29:36.228134] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.228142] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.240167] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.240175] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.252196] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.252204] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.255888] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:28.467 [2024-10-17 17:29:36.264225] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.264235] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.276265] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.276278] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.288291] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.288302] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.300320] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.300331] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.312349] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.312357] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.324381] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.324390] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.336425] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.336442] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.348447] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.348458] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.360481] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.360492] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.372510] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.372519] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.467 [2024-10-17 17:29:36.384540] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.467 [2024-10-17 17:29:36.384548] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.727 [2024-10-17 17:29:36.396571] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.727 [2024-10-17 17:29:36.396579] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.727 [2024-10-17 17:29:36.408603] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.727 [2024-10-17 17:29:36.408613] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.420633] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.420641] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.432666] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.432673] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.444700] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.444708] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.456732] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.456742] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.468762] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.468769] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.480792] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.480800] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.492823] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.492831] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.504854] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.504863] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.516891] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.516907] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 Running I/O for 5 seconds... 00:08:28.728 [2024-10-17 17:29:36.531841] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.531858] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.544931] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.544948] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.558452] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.558469] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.571111] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.571128] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.584317] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.584333] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.597368] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.597383] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.610467] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.610484] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.623802] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.623819] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.728 [2024-10-17 17:29:36.637298] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.728 [2024-10-17 17:29:36.637314] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.650370] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.650385] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.663511] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.663527] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.676783] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.676799] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.689613] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.689630] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.702715] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.702731] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.715781] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.715797] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.728781] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.728797] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.741537] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.741555] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.754770] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.754787] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.988 [2024-10-17 17:29:36.767070] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.988 [2024-10-17 17:29:36.767086] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.779545] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.779561] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.792366] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.792382] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.805802] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.805817] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.818406] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.818422] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.831675] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.831694] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.844787] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.844803] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.857657] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.857673] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.870840] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.870856] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.884679] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.884699] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:28.989 [2024-10-17 17:29:36.897784] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:28.989 [2024-10-17 17:29:36.897799] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:36.911084] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:36.911099] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:36.924640] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:36.924656] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:36.937152] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:36.937167] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:36.950073] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:36.950089] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:36.962591] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:36.962606] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:36.975237] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:36.975256] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:36.988417] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:36.988433] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.001344] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.001360] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.014318] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.014334] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.027442] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.027458] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.041042] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.041057] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.053871] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.053887] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.066970] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.066986] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.079891] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.079906] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.091866] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.091881] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.105572] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.105588] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.118327] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.118343] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.131526] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.131541] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.144652] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.144667] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.249 [2024-10-17 17:29:37.157681] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.249 [2024-10-17 17:29:37.157701] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.170990] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.171005] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.183949] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.183965] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.196855] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.196870] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.210250] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.210265] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.223399] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.223418] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.236496] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.236511] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.249096] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.249112] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.262395] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.262411] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.275136] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.275151] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.288221] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.288236] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.301296] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.301311] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.314169] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.314184] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.326684] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.326704] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.340210] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.340226] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.353906] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.353922] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.367127] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.367143] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.380373] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.380388] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.394018] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.394034] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.406204] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.406220] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.511 [2024-10-17 17:29:37.419788] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.511 [2024-10-17 17:29:37.419803] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.433213] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.433229] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.445712] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.445727] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.459135] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.459150] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.471551] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.471574] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.484954] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.484969] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.498440] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.498455] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.511564] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.511580] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 19175.00 IOPS, 149.80 MiB/s [2024-10-17T15:29:37.691Z] [2024-10-17 17:29:37.524924] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.524939] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.538189] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.538204] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.551814] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.551829] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.565176] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.565191] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.578893] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.578909] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.592446] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.592462] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.605735] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.605750] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.618235] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.618251] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.630661] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.630676] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.644312] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.644327] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.657639] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.657653] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.671038] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.671054] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:29.772 [2024-10-17 17:29:37.684178] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:29.772 [2024-10-17 17:29:37.684193] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.697580] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.697595] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.711172] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.711187] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.724435] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.724451] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.737525] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.737540] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.750516] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.750531] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.763231] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.763245] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.776200] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.776215] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.788770] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.788785] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.800993] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.801008] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.813796] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.813811] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.826270] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.826285] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.839581] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.839596] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.852646] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.852662] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.865447] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.865461] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.878399] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.878415] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.891858] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.891874] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.905210] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.905225] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.918623] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.918638] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.931413] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.931428] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.033 [2024-10-17 17:29:37.944023] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.033 [2024-10-17 17:29:37.944038] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:37.957188] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:37.957203] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:37.970139] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:37.970154] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:37.983671] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:37.983687] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:37.997170] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:37.997186] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.010070] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.010086] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.022488] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.022503] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.034934] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.034949] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.048092] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.048108] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.061486] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.061500] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.074313] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.074328] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.087255] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.087270] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.100114] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.100129] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.113733] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.113748] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.126973] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.126989] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.139348] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.139362] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.152913] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.152928] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.165806] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.165820] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.179506] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.179521] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.192938] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.192953] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.294 [2024-10-17 17:29:38.205856] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.294 [2024-10-17 17:29:38.205872] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.554 [2024-10-17 17:29:38.218565] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.554 [2024-10-17 17:29:38.218581] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.554 [2024-10-17 17:29:38.232274] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.232290] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.244905] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.244920] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.257595] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.257611] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.271162] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.271178] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.284578] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.284593] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.297198] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.297214] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.309535] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.309551] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.322689] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.322707] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.335608] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.335624] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.348723] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.348739] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.362271] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.362287] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.375154] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.375169] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.388522] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.388537] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.401917] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.401932] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.414819] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.414834] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.428080] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.428095] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.441425] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.441440] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.455013] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.455034] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.555 [2024-10-17 17:29:38.467383] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.555 [2024-10-17 17:29:38.467399] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.815 [2024-10-17 17:29:38.480237] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.815 [2024-10-17 17:29:38.480253] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.815 [2024-10-17 17:29:38.493705] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.815 [2024-10-17 17:29:38.493721] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.815 [2024-10-17 17:29:38.506443] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.815 [2024-10-17 17:29:38.506458] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.815 [2024-10-17 17:29:38.519551] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.815 [2024-10-17 17:29:38.519566] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.815 19235.00 IOPS, 150.27 MiB/s [2024-10-17T15:29:38.734Z] [2024-10-17 17:29:38.532594] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.532610] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.546024] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.546039] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.559633] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.559648] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.572924] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.572939] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.586026] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.586041] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.599616] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.599632] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.613267] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.613282] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.626849] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.626864] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.640413] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.640429] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.653145] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.653161] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.665660] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.665675] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.679332] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.679347] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.691817] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.691833] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.704671] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.704702] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.717462] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.717477] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:30.816 [2024-10-17 17:29:38.730137] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:30.816 [2024-10-17 17:29:38.730152] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.076 [2024-10-17 17:29:38.743455] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.743471] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.756603] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.756619] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.769846] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.769862] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.783260] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.783275] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.796487] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.796503] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.809599] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.809615] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.822549] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.822565] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.835772] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.835787] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.848735] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.848750] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.862150] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.862165] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.875427] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.875442] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.887949] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.887964] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.900945] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.900960] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.914638] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.914653] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.927897] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.927912] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.941532] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.941547] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.955003] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.955021] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.967900] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.967915] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.981210] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.981225] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.077 [2024-10-17 17:29:38.994857] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.077 [2024-10-17 17:29:38.994872] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.007298] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.007313] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.019948] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.019963] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.033225] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.033240] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.046441] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.046456] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.060023] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.060038] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.073706] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.073721] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.086732] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.086747] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.100299] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.100314] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.112951] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.112967] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.126306] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.126321] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.139778] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.139793] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.152872] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.152887] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.165809] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.165824] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.178907] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.178921] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.191832] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.191847] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.204917] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.204936] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.217386] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.217402] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.230648] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.230664] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.244121] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.244136] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.338 [2024-10-17 17:29:39.256749] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.338 [2024-10-17 17:29:39.256764] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.269530] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.269545] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.282840] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.282855] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.296076] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.296091] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.309140] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.309154] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.321801] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.321816] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.335088] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.335102] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.347550] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.347565] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.360511] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.360526] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.373851] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.373866] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.386603] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.386618] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.399174] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.399189] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.599 [2024-10-17 17:29:39.411765] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.599 [2024-10-17 17:29:39.411780] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.424046] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.424061] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.437568] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.437584] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.450025] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.450040] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.462713] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.462728] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.475302] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.475317] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.488591] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.488606] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.502174] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.502189] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.600 [2024-10-17 17:29:39.515313] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.600 [2024-10-17 17:29:39.515328] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.528218] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.528233] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 19267.33 IOPS, 150.53 MiB/s [2024-10-17T15:29:39.779Z] [2024-10-17 17:29:39.541799] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.541814] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.554765] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.554780] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.567886] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.567901] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.581237] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.581252] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.594582] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.594598] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.608210] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.608225] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.621237] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.621253] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.860 [2024-10-17 17:29:39.634564] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.860 [2024-10-17 17:29:39.634580] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.647810] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.647826] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.661352] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.661367] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.673736] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.673751] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.686409] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.686424] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.699169] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.699184] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.712631] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.712646] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.725041] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.725056] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.737417] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.737432] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.750660] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.750675] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.764139] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.764154] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:31.861 [2024-10-17 17:29:39.777256] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:31.861 [2024-10-17 17:29:39.777271] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.789827] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.789842] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.803323] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.803338] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.816868] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.816883] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.830188] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.830203] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.843522] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.843537] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.856974] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.856990] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.869506] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.869521] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.882748] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.882763] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.895668] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.895683] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.909048] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.909064] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.922558] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.922574] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.935347] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.935366] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.947879] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.947895] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.961583] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.961598] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.974221] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.974236] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.986496] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.986511] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:39.999756] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:39.999771] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:40.012357] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:40.012373] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:40.024999] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:40.025014] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.122 [2024-10-17 17:29:40.038486] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.122 [2024-10-17 17:29:40.038503] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.051739] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.051758] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.065319] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.065336] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.077779] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.077795] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.090831] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.090847] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.104246] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.104262] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.116798] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.116813] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.129354] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.129369] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.137189] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.137204] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.146126] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.146142] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.154625] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.154641] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.163823] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.163844] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.172976] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.172991] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.181031] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.181047] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.194579] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.194595] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.208090] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.208106] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.220637] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.220653] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.233390] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.233405] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.246253] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.246269] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.259547] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.259562] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.272641] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.272657] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.285153] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.285170] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.383 [2024-10-17 17:29:40.298261] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.383 [2024-10-17 17:29:40.298276] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.310991] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.644 [2024-10-17 17:29:40.311007] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.324171] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.644 [2024-10-17 17:29:40.324187] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.336606] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.644 [2024-10-17 17:29:40.336623] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.349380] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.644 [2024-10-17 17:29:40.349396] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.363448] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.644 [2024-10-17 17:29:40.363465] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.375908] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.644 [2024-10-17 17:29:40.375924] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.388620] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.644 [2024-10-17 17:29:40.388635] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.644 [2024-10-17 17:29:40.402011] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.402030] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.415228] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.415244] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.428835] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.428851] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.442121] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.442137] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.455478] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.455494] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.468596] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.468611] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.481565] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.481581] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.494483] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.494498] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.507984] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.507999] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.521422] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.521438] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 19261.00 IOPS, 150.48 MiB/s [2024-10-17T15:29:40.564Z] [2024-10-17 17:29:40.535043] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.535058] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.548344] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.548360] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.645 [2024-10-17 17:29:40.561508] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.645 [2024-10-17 17:29:40.561524] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.574188] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.574204] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.587875] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.587890] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.601468] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.601483] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.614689] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.614707] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.628278] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.628294] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.641844] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.641859] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.655106] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.655122] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.667626] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.667640] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.681146] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.681162] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.694139] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.694154] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.707091] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.707106] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.720777] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.720793] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.733098] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.733114] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.745885] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.745900] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.759111] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.759127] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.771890] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.771904] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.785217] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.785232] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.798477] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.798491] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.811728] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.811743] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:32.906 [2024-10-17 17:29:40.824749] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:32.906 [2024-10-17 17:29:40.824765] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.838276] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.838292] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.851511] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.851526] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.864872] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.864888] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.878275] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.878290] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.891787] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.891803] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.905267] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.905283] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.918507] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.918522] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.932031] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.932045] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.944968] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.944982] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.958471] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.958486] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.971959] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.971974] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.984858] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.984873] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:40.998264] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:40.998279] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:41.011697] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:41.011712] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:41.025249] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:41.025264] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:41.038220] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:41.038235] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:41.051457] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:41.051472] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:41.064927] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:41.064941] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.167 [2024-10-17 17:29:41.078105] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.167 [2024-10-17 17:29:41.078120] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.091470] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.091485] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.105094] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.105110] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.118215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.118230] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.131714] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.131729] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.145105] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.145119] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.158696] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.158712] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.171983] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.171997] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.185219] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.185234] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.198597] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.198613] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.211618] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.211633] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.224583] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.224598] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.237223] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.237238] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.249941] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.249956] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.263704] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.263719] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.276911] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.276925] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.290300] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.290315] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.303096] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.303112] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.315861] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.315875] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.328985] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.329000] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.428 [2024-10-17 17:29:41.342300] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.428 [2024-10-17 17:29:41.342315] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.355146] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.355161] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.368193] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.368209] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.381581] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.381596] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.395351] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.395366] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.408416] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.408431] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.421566] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.421581] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.435041] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.435056] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.448380] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.448395] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.461740] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.461755] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.475236] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.475251] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.488550] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.488564] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.501661] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.501675] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.514261] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.514276] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.527611] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.527626] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 19276.00 IOPS, 150.59 MiB/s [2024-10-17T15:29:41.608Z] [2024-10-17 17:29:41.538907] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.538922] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 00:08:33.689 Latency(us) 00:08:33.689 [2024-10-17T15:29:41.608Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:08:33.689 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:08:33.689 Nvme1n1 : 5.01 19276.61 150.60 0.00 0.00 6634.22 2921.81 15291.73 00:08:33.689 [2024-10-17T15:29:41.608Z] =================================================================================================================== 00:08:33.689 [2024-10-17T15:29:41.608Z] Total : 19276.61 150.60 0.00 0.00 6634.22 2921.81 15291.73 00:08:33.689 [2024-10-17 17:29:41.549522] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.549534] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.561563] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.561578] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.573586] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.573598] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.585618] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.585629] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.689 [2024-10-17 17:29:41.597645] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.689 [2024-10-17 17:29:41.597659] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.951 [2024-10-17 17:29:41.609674] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.951 [2024-10-17 17:29:41.609682] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.951 [2024-10-17 17:29:41.621709] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.951 [2024-10-17 17:29:41.621719] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.951 [2024-10-17 17:29:41.633740] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:08:33.951 [2024-10-17 17:29:41.633750] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:33.951 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (4085640) - No such process 00:08:33.951 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 4085640 00:08:33.951 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:08:33.951 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.951 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:33.951 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:33.952 delay0 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.952 17:29:41 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:08:33.952 [2024-10-17 17:29:41.753903] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:08:40.537 Initializing NVMe Controllers 00:08:40.537 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:08:40.537 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:08:40.537 Initialization complete. Launching workers. 00:08:40.537 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 724 00:08:40.537 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 1014, failed to submit 30 00:08:40.537 success 813, unsuccessful 201, failed 0 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@514 -- # nvmfcleanup 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:40.537 rmmod nvme_tcp 00:08:40.537 rmmod nvme_fabrics 00:08:40.537 rmmod nvme_keyring 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@515 -- # '[' -n 4083291 ']' 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@516 -- # killprocess 4083291 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@950 -- # '[' -z 4083291 ']' 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@954 -- # kill -0 4083291 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@955 -- # uname 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:40.537 17:29:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4083291 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4083291' 00:08:40.537 killing process with pid 4083291 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@969 -- # kill 4083291 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@974 -- # wait 4083291 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-save 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-restore 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:40.537 17:29:48 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:42.447 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:08:42.447 00:08:42.447 real 0m33.576s 00:08:42.447 user 0m44.456s 00:08:42.447 sys 0m10.940s 00:08:42.447 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:42.447 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:08:42.447 ************************************ 00:08:42.447 END TEST nvmf_zcopy 00:08:42.447 ************************************ 00:08:42.447 17:29:50 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:08:42.447 17:29:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:08:42.448 17:29:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:42.448 17:29:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:08:42.448 ************************************ 00:08:42.448 START TEST nvmf_nmic 00:08:42.448 ************************************ 00:08:42.448 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:08:42.709 * Looking for test storage... 00:08:42.709 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1691 -- # lcov --version 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:08:42.709 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:42.709 --rc genhtml_branch_coverage=1 00:08:42.709 --rc genhtml_function_coverage=1 00:08:42.709 --rc genhtml_legend=1 00:08:42.709 --rc geninfo_all_blocks=1 00:08:42.709 --rc geninfo_unexecuted_blocks=1 00:08:42.709 00:08:42.709 ' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:08:42.709 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:42.709 --rc genhtml_branch_coverage=1 00:08:42.709 --rc genhtml_function_coverage=1 00:08:42.709 --rc genhtml_legend=1 00:08:42.709 --rc geninfo_all_blocks=1 00:08:42.709 --rc geninfo_unexecuted_blocks=1 00:08:42.709 00:08:42.709 ' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:08:42.709 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:42.709 --rc genhtml_branch_coverage=1 00:08:42.709 --rc genhtml_function_coverage=1 00:08:42.709 --rc genhtml_legend=1 00:08:42.709 --rc geninfo_all_blocks=1 00:08:42.709 --rc geninfo_unexecuted_blocks=1 00:08:42.709 00:08:42.709 ' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:08:42.709 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:42.709 --rc genhtml_branch_coverage=1 00:08:42.709 --rc genhtml_function_coverage=1 00:08:42.709 --rc genhtml_legend=1 00:08:42.709 --rc geninfo_all_blocks=1 00:08:42.709 --rc geninfo_unexecuted_blocks=1 00:08:42.709 00:08:42.709 ' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:42.709 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:42.710 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@474 -- # prepare_net_devs 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@436 -- # local -g is_hw=no 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@438 -- # remove_spdk_ns 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:08:42.710 17:29:50 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:50.851 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:08:50.852 Found 0000:31:00.0 (0x8086 - 0x159b) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:08:50.852 Found 0000:31:00.1 (0x8086 - 0x159b) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:08:50.852 Found net devices under 0000:31:00.0: cvl_0_0 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:08:50.852 Found net devices under 0000:31:00.1: cvl_0_1 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # is_hw=yes 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:08:50.852 17:29:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:08:50.852 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:50.852 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.625 ms 00:08:50.852 00:08:50.852 --- 10.0.0.2 ping statistics --- 00:08:50.852 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:50.852 rtt min/avg/max/mdev = 0.625/0.625/0.625/0.000 ms 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:50.852 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:50.852 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.277 ms 00:08:50.852 00:08:50.852 --- 10.0.0.1 ping statistics --- 00:08:50.852 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:50.852 rtt min/avg/max/mdev = 0.277/0.277/0.277/0.000 ms 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@448 -- # return 0 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@507 -- # nvmfpid=4092353 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@508 -- # waitforlisten 4092353 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@831 -- # '[' -z 4092353 ']' 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:50.852 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:50.852 17:29:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:50.852 [2024-10-17 17:29:58.191558] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:08:50.852 [2024-10-17 17:29:58.191627] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:50.853 [2024-10-17 17:29:58.282482] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:50.853 [2024-10-17 17:29:58.338839] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:50.853 [2024-10-17 17:29:58.338897] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:50.853 [2024-10-17 17:29:58.338905] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:50.853 [2024-10-17 17:29:58.338912] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:50.853 [2024-10-17 17:29:58.338918] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:50.853 [2024-10-17 17:29:58.341353] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:50.853 [2024-10-17 17:29:58.341511] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:50.853 [2024-10-17 17:29:58.341671] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:08:50.853 [2024-10-17 17:29:58.341672] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:51.114 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:51.114 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@864 -- # return 0 00:08:51.114 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:08:51.114 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:51.114 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 [2024-10-17 17:29:59.062184] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 Malloc0 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 [2024-10-17 17:29:59.143077] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:08:51.376 test case1: single bdev can't be used in multiple subsystems 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 [2024-10-17 17:29:59.178853] bdev.c:8202:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:08:51.376 [2024-10-17 17:29:59.178887] subsystem.c:2157:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:08:51.376 [2024-10-17 17:29:59.178895] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:08:51.376 request: 00:08:51.376 { 00:08:51.376 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:08:51.376 "namespace": { 00:08:51.376 "bdev_name": "Malloc0", 00:08:51.376 "no_auto_visible": false 00:08:51.376 }, 00:08:51.376 "method": "nvmf_subsystem_add_ns", 00:08:51.376 "req_id": 1 00:08:51.376 } 00:08:51.376 Got JSON-RPC error response 00:08:51.376 response: 00:08:51.376 { 00:08:51.376 "code": -32602, 00:08:51.376 "message": "Invalid parameters" 00:08:51.376 } 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:08:51.376 Adding namespace failed - expected result. 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:08:51.376 test case2: host connect to nvmf target in multiple paths 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:08:51.376 [2024-10-17 17:29:59.191092] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.376 17:29:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:53.290 17:30:00 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:08:54.668 17:30:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:08:54.668 17:30:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1198 -- # local i=0 00:08:54.668 17:30:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:54.668 17:30:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:54.668 17:30:02 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1205 -- # sleep 2 00:08:56.579 17:30:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:56.579 17:30:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:56.579 17:30:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:56.579 17:30:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:56.579 17:30:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:56.579 17:30:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1208 -- # return 0 00:08:56.579 17:30:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:08:56.579 [global] 00:08:56.579 thread=1 00:08:56.579 invalidate=1 00:08:56.579 rw=write 00:08:56.579 time_based=1 00:08:56.579 runtime=1 00:08:56.579 ioengine=libaio 00:08:56.579 direct=1 00:08:56.579 bs=4096 00:08:56.579 iodepth=1 00:08:56.579 norandommap=0 00:08:56.579 numjobs=1 00:08:56.579 00:08:56.579 verify_dump=1 00:08:56.579 verify_backlog=512 00:08:56.579 verify_state_save=0 00:08:56.579 do_verify=1 00:08:56.579 verify=crc32c-intel 00:08:56.579 [job0] 00:08:56.579 filename=/dev/nvme0n1 00:08:56.579 Could not set queue depth (nvme0n1) 00:08:56.838 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:08:56.838 fio-3.35 00:08:56.838 Starting 1 thread 00:08:58.219 00:08:58.219 job0: (groupid=0, jobs=1): err= 0: pid=4093958: Thu Oct 17 17:30:05 2024 00:08:58.219 read: IOPS=18, BW=73.3KiB/s (75.0kB/s)(76.0KiB/1037msec) 00:08:58.219 slat (nsec): min=10419, max=27734, avg=26153.21, stdev=3815.44 00:08:58.219 clat (usec): min=40898, max=41465, avg=40987.32, stdev=118.40 00:08:58.219 lat (usec): min=40925, max=41476, avg=41013.47, stdev=114.67 00:08:58.219 clat percentiles (usec): 00:08:58.219 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:08:58.219 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:08:58.219 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41681], 00:08:58.219 | 99.00th=[41681], 99.50th=[41681], 99.90th=[41681], 99.95th=[41681], 00:08:58.219 | 99.99th=[41681] 00:08:58.219 write: IOPS=493, BW=1975KiB/s (2022kB/s)(2048KiB/1037msec); 0 zone resets 00:08:58.219 slat (usec): min=10, max=26907, avg=79.30, stdev=1188.05 00:08:58.220 clat (usec): min=240, max=582, avg=415.75, stdev=65.70 00:08:58.220 lat (usec): min=251, max=27257, avg=495.05, stdev=1187.29 00:08:58.220 clat percentiles (usec): 00:08:58.220 | 1.00th=[ 258], 5.00th=[ 289], 10.00th=[ 334], 20.00th=[ 351], 00:08:58.220 | 30.00th=[ 371], 40.00th=[ 408], 50.00th=[ 433], 60.00th=[ 453], 00:08:58.220 | 70.00th=[ 465], 80.00th=[ 474], 90.00th=[ 486], 95.00th=[ 494], 00:08:58.220 | 99.00th=[ 519], 99.50th=[ 553], 99.90th=[ 586], 99.95th=[ 586], 00:08:58.220 | 99.99th=[ 586] 00:08:58.220 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:08:58.220 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:08:58.220 lat (usec) : 250=0.19%, 500=93.03%, 750=3.20% 00:08:58.220 lat (msec) : 50=3.58% 00:08:58.220 cpu : usr=0.48%, sys=1.45%, ctx=535, majf=0, minf=1 00:08:58.220 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:08:58.220 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:08:58.220 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:08:58.220 issued rwts: total=19,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:08:58.220 latency : target=0, window=0, percentile=100.00%, depth=1 00:08:58.220 00:08:58.220 Run status group 0 (all jobs): 00:08:58.220 READ: bw=73.3KiB/s (75.0kB/s), 73.3KiB/s-73.3KiB/s (75.0kB/s-75.0kB/s), io=76.0KiB (77.8kB), run=1037-1037msec 00:08:58.220 WRITE: bw=1975KiB/s (2022kB/s), 1975KiB/s-1975KiB/s (2022kB/s-2022kB/s), io=2048KiB (2097kB), run=1037-1037msec 00:08:58.220 00:08:58.220 Disk stats (read/write): 00:08:58.220 nvme0n1: ios=40/512, merge=0/0, ticks=1579/216, in_queue=1795, util=98.60% 00:08:58.220 17:30:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:58.480 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1219 -- # local i=0 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1231 -- # return 0 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@514 -- # nvmfcleanup 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:08:58.480 rmmod nvme_tcp 00:08:58.480 rmmod nvme_fabrics 00:08:58.480 rmmod nvme_keyring 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@515 -- # '[' -n 4092353 ']' 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@516 -- # killprocess 4092353 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@950 -- # '[' -z 4092353 ']' 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@954 -- # kill -0 4092353 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@955 -- # uname 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4092353 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4092353' 00:08:58.480 killing process with pid 4092353 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@969 -- # kill 4092353 00:08:58.480 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@974 -- # wait 4092353 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-save 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-restore 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:58.740 17:30:06 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:00.649 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:09:00.649 00:09:00.649 real 0m18.250s 00:09:00.649 user 0m49.215s 00:09:00.649 sys 0m6.644s 00:09:00.649 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:00.649 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:09:00.649 ************************************ 00:09:00.649 END TEST nvmf_nmic 00:09:00.649 ************************************ 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:09:00.910 ************************************ 00:09:00.910 START TEST nvmf_fio_target 00:09:00.910 ************************************ 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:09:00.910 * Looking for test storage... 00:09:00.910 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1691 -- # lcov --version 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:00.910 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:09:00.910 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:00.910 --rc genhtml_branch_coverage=1 00:09:00.910 --rc genhtml_function_coverage=1 00:09:00.911 --rc genhtml_legend=1 00:09:00.911 --rc geninfo_all_blocks=1 00:09:00.911 --rc geninfo_unexecuted_blocks=1 00:09:00.911 00:09:00.911 ' 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:09:00.911 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:00.911 --rc genhtml_branch_coverage=1 00:09:00.911 --rc genhtml_function_coverage=1 00:09:00.911 --rc genhtml_legend=1 00:09:00.911 --rc geninfo_all_blocks=1 00:09:00.911 --rc geninfo_unexecuted_blocks=1 00:09:00.911 00:09:00.911 ' 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:09:00.911 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:00.911 --rc genhtml_branch_coverage=1 00:09:00.911 --rc genhtml_function_coverage=1 00:09:00.911 --rc genhtml_legend=1 00:09:00.911 --rc geninfo_all_blocks=1 00:09:00.911 --rc geninfo_unexecuted_blocks=1 00:09:00.911 00:09:00.911 ' 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:09:00.911 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:00.911 --rc genhtml_branch_coverage=1 00:09:00.911 --rc genhtml_function_coverage=1 00:09:00.911 --rc genhtml_legend=1 00:09:00.911 --rc geninfo_all_blocks=1 00:09:00.911 --rc geninfo_unexecuted_blocks=1 00:09:00.911 00:09:00.911 ' 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:00.911 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:01.171 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:01.172 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:09:01.172 17:30:08 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:09:09.466 Found 0000:31:00.0 (0x8086 - 0x159b) 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:09.466 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:09:09.466 Found 0000:31:00.1 (0x8086 - 0x159b) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:09:09.467 Found net devices under 0000:31:00.0: cvl_0_0 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:09:09.467 Found net devices under 0000:31:00.1: cvl_0_1 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # is_hw=yes 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:09.467 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:09.467 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.678 ms 00:09:09.467 00:09:09.467 --- 10.0.0.2 ping statistics --- 00:09:09.467 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:09.467 rtt min/avg/max/mdev = 0.678/0.678/0.678/0.000 ms 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:09.467 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:09.467 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.279 ms 00:09:09.467 00:09:09.467 --- 10.0.0.1 ping statistics --- 00:09:09.467 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:09.467 rtt min/avg/max/mdev = 0.279/0.279/0.279/0.000 ms 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@448 -- # return 0 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@507 -- # nvmfpid=4098901 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@508 -- # waitforlisten 4098901 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@831 -- # '[' -z 4098901 ']' 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:09.467 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:09.467 17:30:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:09.467 [2024-10-17 17:30:16.479442] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:09:09.467 [2024-10-17 17:30:16.479504] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:09.467 [2024-10-17 17:30:16.569073] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:09.467 [2024-10-17 17:30:16.623226] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:09.467 [2024-10-17 17:30:16.623281] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:09.467 [2024-10-17 17:30:16.623290] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:09.467 [2024-10-17 17:30:16.623297] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:09.467 [2024-10-17 17:30:16.623304] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:09.467 [2024-10-17 17:30:16.625443] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:09.467 [2024-10-17 17:30:16.625603] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:09.467 [2024-10-17 17:30:16.625754] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:09.467 [2024-10-17 17:30:16.625753] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:09.467 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:09.467 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@864 -- # return 0 00:09:09.467 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:09:09.467 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:09:09.467 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:09.467 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:09.467 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:09:09.727 [2024-10-17 17:30:17.522900] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:09.727 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:09:09.987 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:09:09.987 17:30:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:09:10.247 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:09:10.247 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:09:10.508 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:09:10.508 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:09:10.768 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:09:10.768 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:09:10.768 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:09:11.027 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:09:11.027 17:30:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:09:11.286 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:09:11.286 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:09:11.545 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:09:11.545 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:09:11.545 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:09:11.805 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:09:11.805 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:09:12.066 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:09:12.066 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:09:12.066 17:30:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:12.326 [2024-10-17 17:30:20.137526] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:12.326 17:30:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:09:12.587 17:30:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:09:12.846 17:30:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:09:14.229 17:30:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:09:14.229 17:30:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1198 -- # local i=0 00:09:14.229 17:30:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:09:14.229 17:30:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1200 -- # [[ -n 4 ]] 00:09:14.229 17:30:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1201 -- # nvme_device_counter=4 00:09:14.229 17:30:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1205 -- # sleep 2 00:09:16.771 17:30:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:09:16.771 17:30:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:09:16.771 17:30:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:09:16.771 17:30:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1207 -- # nvme_devices=4 00:09:16.771 17:30:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:09:16.771 17:30:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1208 -- # return 0 00:09:16.771 17:30:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:09:16.771 [global] 00:09:16.771 thread=1 00:09:16.771 invalidate=1 00:09:16.771 rw=write 00:09:16.771 time_based=1 00:09:16.771 runtime=1 00:09:16.771 ioengine=libaio 00:09:16.771 direct=1 00:09:16.771 bs=4096 00:09:16.771 iodepth=1 00:09:16.771 norandommap=0 00:09:16.771 numjobs=1 00:09:16.771 00:09:16.771 verify_dump=1 00:09:16.771 verify_backlog=512 00:09:16.771 verify_state_save=0 00:09:16.771 do_verify=1 00:09:16.771 verify=crc32c-intel 00:09:16.772 [job0] 00:09:16.772 filename=/dev/nvme0n1 00:09:16.772 [job1] 00:09:16.772 filename=/dev/nvme0n2 00:09:16.772 [job2] 00:09:16.772 filename=/dev/nvme0n3 00:09:16.772 [job3] 00:09:16.772 filename=/dev/nvme0n4 00:09:16.772 Could not set queue depth (nvme0n1) 00:09:16.772 Could not set queue depth (nvme0n2) 00:09:16.772 Could not set queue depth (nvme0n3) 00:09:16.772 Could not set queue depth (nvme0n4) 00:09:16.772 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:16.772 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:16.772 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:16.772 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:16.772 fio-3.35 00:09:16.772 Starting 4 threads 00:09:18.157 00:09:18.157 job0: (groupid=0, jobs=1): err= 0: pid=4100805: Thu Oct 17 17:30:25 2024 00:09:18.157 read: IOPS=571, BW=2286KiB/s (2341kB/s)(2288KiB/1001msec) 00:09:18.157 slat (nsec): min=7217, max=61235, avg=25275.05, stdev=5058.29 00:09:18.157 clat (usec): min=436, max=1067, avg=754.17, stdev=134.98 00:09:18.157 lat (usec): min=462, max=1092, avg=779.45, stdev=134.96 00:09:18.157 clat percentiles (usec): 00:09:18.157 | 1.00th=[ 465], 5.00th=[ 529], 10.00th=[ 553], 20.00th=[ 611], 00:09:18.157 | 30.00th=[ 676], 40.00th=[ 717], 50.00th=[ 766], 60.00th=[ 816], 00:09:18.157 | 70.00th=[ 857], 80.00th=[ 881], 90.00th=[ 914], 95.00th=[ 947], 00:09:18.157 | 99.00th=[ 1004], 99.50th=[ 1029], 99.90th=[ 1074], 99.95th=[ 1074], 00:09:18.157 | 99.99th=[ 1074] 00:09:18.157 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:09:18.157 slat (usec): min=9, max=6384, avg=38.89, stdev=198.65 00:09:18.157 clat (usec): min=130, max=819, avg=489.98, stdev=110.06 00:09:18.157 lat (usec): min=140, max=7080, avg=528.86, stdev=233.72 00:09:18.157 clat percentiles (usec): 00:09:18.157 | 1.00th=[ 245], 5.00th=[ 302], 10.00th=[ 355], 20.00th=[ 388], 00:09:18.157 | 30.00th=[ 424], 40.00th=[ 469], 50.00th=[ 494], 60.00th=[ 523], 00:09:18.157 | 70.00th=[ 545], 80.00th=[ 594], 90.00th=[ 627], 95.00th=[ 660], 00:09:18.157 | 99.00th=[ 725], 99.50th=[ 734], 99.90th=[ 783], 99.95th=[ 824], 00:09:18.157 | 99.99th=[ 824] 00:09:18.157 bw ( KiB/s): min= 4096, max= 4096, per=31.10%, avg=4096.00, stdev= 0.00, samples=1 00:09:18.157 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:18.157 lat (usec) : 250=0.88%, 500=33.46%, 750=46.18%, 1000=19.11% 00:09:18.157 lat (msec) : 2=0.38% 00:09:18.157 cpu : usr=2.80%, sys=4.50%, ctx=1600, majf=0, minf=1 00:09:18.157 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:18.157 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.157 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.157 issued rwts: total=572,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:18.157 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:18.157 job1: (groupid=0, jobs=1): err= 0: pid=4100807: Thu Oct 17 17:30:25 2024 00:09:18.157 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:09:18.157 slat (nsec): min=7884, max=84463, avg=25133.08, stdev=3568.37 00:09:18.157 clat (usec): min=545, max=1370, avg=969.27, stdev=107.17 00:09:18.157 lat (usec): min=571, max=1395, avg=994.40, stdev=107.06 00:09:18.157 clat percentiles (usec): 00:09:18.157 | 1.00th=[ 693], 5.00th=[ 783], 10.00th=[ 816], 20.00th=[ 881], 00:09:18.157 | 30.00th=[ 930], 40.00th=[ 963], 50.00th=[ 979], 60.00th=[ 1004], 00:09:18.157 | 70.00th=[ 1029], 80.00th=[ 1057], 90.00th=[ 1090], 95.00th=[ 1123], 00:09:18.157 | 99.00th=[ 1205], 99.50th=[ 1254], 99.90th=[ 1369], 99.95th=[ 1369], 00:09:18.157 | 99.99th=[ 1369] 00:09:18.157 write: IOPS=735, BW=2941KiB/s (3012kB/s)(2944KiB/1001msec); 0 zone resets 00:09:18.157 slat (nsec): min=9668, max=65305, avg=29044.96, stdev=8889.53 00:09:18.157 clat (usec): min=174, max=1189, avg=624.78, stdev=123.54 00:09:18.157 lat (usec): min=184, max=1220, avg=653.83, stdev=127.18 00:09:18.157 clat percentiles (usec): 00:09:18.157 | 1.00th=[ 338], 5.00th=[ 383], 10.00th=[ 457], 20.00th=[ 537], 00:09:18.157 | 30.00th=[ 570], 40.00th=[ 603], 50.00th=[ 635], 60.00th=[ 668], 00:09:18.157 | 70.00th=[ 693], 80.00th=[ 725], 90.00th=[ 758], 95.00th=[ 799], 00:09:18.157 | 99.00th=[ 889], 99.50th=[ 938], 99.90th=[ 1188], 99.95th=[ 1188], 00:09:18.157 | 99.99th=[ 1188] 00:09:18.157 bw ( KiB/s): min= 4096, max= 4096, per=31.10%, avg=4096.00, stdev= 0.00, samples=1 00:09:18.157 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:18.157 lat (usec) : 250=0.16%, 500=8.73%, 750=43.75%, 1000=29.73% 00:09:18.157 lat (msec) : 2=17.63% 00:09:18.157 cpu : usr=2.00%, sys=3.40%, ctx=1249, majf=0, minf=1 00:09:18.157 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:18.157 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.157 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.157 issued rwts: total=512,736,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:18.157 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:18.157 job2: (groupid=0, jobs=1): err= 0: pid=4100808: Thu Oct 17 17:30:25 2024 00:09:18.157 read: IOPS=84, BW=340KiB/s (348kB/s)(340KiB/1001msec) 00:09:18.157 slat (nsec): min=7324, max=46123, avg=26474.19, stdev=4098.74 00:09:18.157 clat (usec): min=817, max=43275, avg=7757.58, stdev=15302.16 00:09:18.157 lat (usec): min=845, max=43302, avg=7784.06, stdev=15302.04 00:09:18.157 clat percentiles (usec): 00:09:18.157 | 1.00th=[ 816], 5.00th=[ 865], 10.00th=[ 906], 20.00th=[ 955], 00:09:18.157 | 30.00th=[ 988], 40.00th=[ 1004], 50.00th=[ 1029], 60.00th=[ 1045], 00:09:18.158 | 70.00th=[ 1057], 80.00th=[ 1106], 90.00th=[41681], 95.00th=[42206], 00:09:18.158 | 99.00th=[43254], 99.50th=[43254], 99.90th=[43254], 99.95th=[43254], 00:09:18.158 | 99.99th=[43254] 00:09:18.158 write: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec); 0 zone resets 00:09:18.158 slat (nsec): min=9415, max=70184, avg=31496.48, stdev=9062.78 00:09:18.158 clat (usec): min=217, max=1165, avg=622.95, stdev=133.35 00:09:18.158 lat (usec): min=251, max=1200, avg=654.44, stdev=136.40 00:09:18.158 clat percentiles (usec): 00:09:18.158 | 1.00th=[ 338], 5.00th=[ 408], 10.00th=[ 445], 20.00th=[ 506], 00:09:18.158 | 30.00th=[ 553], 40.00th=[ 586], 50.00th=[ 627], 60.00th=[ 668], 00:09:18.158 | 70.00th=[ 693], 80.00th=[ 734], 90.00th=[ 783], 95.00th=[ 832], 00:09:18.158 | 99.00th=[ 922], 99.50th=[ 988], 99.90th=[ 1172], 99.95th=[ 1172], 00:09:18.158 | 99.99th=[ 1172] 00:09:18.158 bw ( KiB/s): min= 4096, max= 4096, per=31.10%, avg=4096.00, stdev= 0.00, samples=1 00:09:18.158 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:18.158 lat (usec) : 250=0.17%, 500=16.42%, 750=55.11%, 1000=19.43% 00:09:18.158 lat (msec) : 2=6.53%, 50=2.35% 00:09:18.158 cpu : usr=1.40%, sys=2.20%, ctx=597, majf=0, minf=1 00:09:18.158 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:18.158 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.158 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.158 issued rwts: total=85,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:18.158 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:18.158 job3: (groupid=0, jobs=1): err= 0: pid=4100809: Thu Oct 17 17:30:25 2024 00:09:18.158 read: IOPS=557, BW=2230KiB/s (2283kB/s)(2232KiB/1001msec) 00:09:18.158 slat (nsec): min=6806, max=48327, avg=25761.33, stdev=6673.88 00:09:18.158 clat (usec): min=325, max=1083, avg=755.40, stdev=119.20 00:09:18.158 lat (usec): min=352, max=1129, avg=781.17, stdev=119.67 00:09:18.158 clat percentiles (usec): 00:09:18.158 | 1.00th=[ 465], 5.00th=[ 545], 10.00th=[ 594], 20.00th=[ 660], 00:09:18.158 | 30.00th=[ 693], 40.00th=[ 725], 50.00th=[ 766], 60.00th=[ 799], 00:09:18.158 | 70.00th=[ 832], 80.00th=[ 865], 90.00th=[ 898], 95.00th=[ 922], 00:09:18.158 | 99.00th=[ 988], 99.50th=[ 1004], 99.90th=[ 1090], 99.95th=[ 1090], 00:09:18.158 | 99.99th=[ 1090] 00:09:18.158 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:09:18.158 slat (nsec): min=9181, max=73011, avg=32132.08, stdev=9037.63 00:09:18.158 clat (usec): min=123, max=804, avg=507.79, stdev=121.42 00:09:18.158 lat (usec): min=135, max=838, avg=539.92, stdev=125.51 00:09:18.158 clat percentiles (usec): 00:09:18.158 | 1.00th=[ 206], 5.00th=[ 289], 10.00th=[ 351], 20.00th=[ 396], 00:09:18.158 | 30.00th=[ 445], 40.00th=[ 486], 50.00th=[ 515], 60.00th=[ 553], 00:09:18.158 | 70.00th=[ 586], 80.00th=[ 619], 90.00th=[ 660], 95.00th=[ 685], 00:09:18.158 | 99.00th=[ 734], 99.50th=[ 742], 99.90th=[ 783], 99.95th=[ 807], 00:09:18.158 | 99.99th=[ 807] 00:09:18.158 bw ( KiB/s): min= 4096, max= 4096, per=31.10%, avg=4096.00, stdev= 0.00, samples=1 00:09:18.158 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:18.158 lat (usec) : 250=1.20%, 500=28.89%, 750=50.76%, 1000=18.96% 00:09:18.158 lat (msec) : 2=0.19% 00:09:18.158 cpu : usr=4.10%, sys=5.40%, ctx=1583, majf=0, minf=1 00:09:18.158 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:18.158 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.158 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:18.158 issued rwts: total=558,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:18.158 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:18.158 00:09:18.158 Run status group 0 (all jobs): 00:09:18.158 READ: bw=6901KiB/s (7067kB/s), 340KiB/s-2286KiB/s (348kB/s-2341kB/s), io=6908KiB (7074kB), run=1001-1001msec 00:09:18.158 WRITE: bw=12.9MiB/s (13.5MB/s), 2046KiB/s-4092KiB/s (2095kB/s-4190kB/s), io=12.9MiB (13.5MB), run=1001-1001msec 00:09:18.158 00:09:18.158 Disk stats (read/write): 00:09:18.158 nvme0n1: ios=534/704, merge=0/0, ticks=1214/317, in_queue=1531, util=86.97% 00:09:18.158 nvme0n2: ios=499/512, merge=0/0, ticks=535/295, in_queue=830, util=86.34% 00:09:18.158 nvme0n3: ios=67/512, merge=0/0, ticks=456/246, in_queue=702, util=88.66% 00:09:18.158 nvme0n4: ios=569/700, merge=0/0, ticks=442/246, in_queue=688, util=95.05% 00:09:18.158 17:30:25 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:09:18.158 [global] 00:09:18.158 thread=1 00:09:18.158 invalidate=1 00:09:18.158 rw=randwrite 00:09:18.158 time_based=1 00:09:18.158 runtime=1 00:09:18.158 ioengine=libaio 00:09:18.158 direct=1 00:09:18.158 bs=4096 00:09:18.158 iodepth=1 00:09:18.158 norandommap=0 00:09:18.158 numjobs=1 00:09:18.158 00:09:18.158 verify_dump=1 00:09:18.158 verify_backlog=512 00:09:18.158 verify_state_save=0 00:09:18.158 do_verify=1 00:09:18.158 verify=crc32c-intel 00:09:18.158 [job0] 00:09:18.158 filename=/dev/nvme0n1 00:09:18.158 [job1] 00:09:18.158 filename=/dev/nvme0n2 00:09:18.158 [job2] 00:09:18.158 filename=/dev/nvme0n3 00:09:18.158 [job3] 00:09:18.158 filename=/dev/nvme0n4 00:09:18.158 Could not set queue depth (nvme0n1) 00:09:18.158 Could not set queue depth (nvme0n2) 00:09:18.158 Could not set queue depth (nvme0n3) 00:09:18.158 Could not set queue depth (nvme0n4) 00:09:18.418 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:18.418 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:18.418 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:18.418 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:18.418 fio-3.35 00:09:18.418 Starting 4 threads 00:09:19.803 00:09:19.803 job0: (groupid=0, jobs=1): err= 0: pid=4101325: Thu Oct 17 17:30:27 2024 00:09:19.803 read: IOPS=16, BW=66.4KiB/s (68.0kB/s)(68.0KiB/1024msec) 00:09:19.803 slat (nsec): min=25565, max=26754, avg=25916.59, stdev=320.71 00:09:19.803 clat (usec): min=40948, max=42055, avg=41580.85, stdev=483.74 00:09:19.803 lat (usec): min=40975, max=42081, avg=41606.76, stdev=483.67 00:09:19.803 clat percentiles (usec): 00:09:19.803 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:09:19.803 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41681], 60.00th=[42206], 00:09:19.803 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:09:19.803 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:09:19.803 | 99.99th=[42206] 00:09:19.803 write: IOPS=500, BW=2000KiB/s (2048kB/s)(2048KiB/1024msec); 0 zone resets 00:09:19.804 slat (nsec): min=9883, max=51120, avg=31984.04, stdev=6113.69 00:09:19.804 clat (usec): min=193, max=944, avg=576.86, stdev=137.16 00:09:19.804 lat (usec): min=226, max=976, avg=608.84, stdev=138.05 00:09:19.804 clat percentiles (usec): 00:09:19.804 | 1.00th=[ 262], 5.00th=[ 326], 10.00th=[ 383], 20.00th=[ 465], 00:09:19.804 | 30.00th=[ 519], 40.00th=[ 553], 50.00th=[ 594], 60.00th=[ 627], 00:09:19.804 | 70.00th=[ 660], 80.00th=[ 693], 90.00th=[ 742], 95.00th=[ 783], 00:09:19.804 | 99.00th=[ 857], 99.50th=[ 873], 99.90th=[ 947], 99.95th=[ 947], 00:09:19.804 | 99.99th=[ 947] 00:09:19.804 bw ( KiB/s): min= 4096, max= 4096, per=40.30%, avg=4096.00, stdev= 0.00, samples=1 00:09:19.804 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:19.804 lat (usec) : 250=0.76%, 500=24.20%, 750=64.27%, 1000=7.56% 00:09:19.804 lat (msec) : 50=3.21% 00:09:19.804 cpu : usr=0.88%, sys=1.56%, ctx=532, majf=0, minf=1 00:09:19.804 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:19.804 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 issued rwts: total=17,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:19.804 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:19.804 job1: (groupid=0, jobs=1): err= 0: pid=4101326: Thu Oct 17 17:30:27 2024 00:09:19.804 read: IOPS=158, BW=634KiB/s (649kB/s)(640KiB/1010msec) 00:09:19.804 slat (nsec): min=9750, max=45546, avg=27981.57, stdev=2372.48 00:09:19.804 clat (usec): min=552, max=42029, avg=4271.87, stdev=11214.39 00:09:19.804 lat (usec): min=581, max=42056, avg=4299.85, stdev=11214.41 00:09:19.804 clat percentiles (usec): 00:09:19.804 | 1.00th=[ 553], 5.00th=[ 693], 10.00th=[ 791], 20.00th=[ 898], 00:09:19.804 | 30.00th=[ 930], 40.00th=[ 955], 50.00th=[ 971], 60.00th=[ 996], 00:09:19.804 | 70.00th=[ 1012], 80.00th=[ 1057], 90.00th=[ 1188], 95.00th=[41681], 00:09:19.804 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:09:19.804 | 99.99th=[42206] 00:09:19.804 write: IOPS=506, BW=2028KiB/s (2076kB/s)(2048KiB/1010msec); 0 zone resets 00:09:19.804 slat (nsec): min=9819, max=55041, avg=30991.46, stdev=8678.18 00:09:19.804 clat (usec): min=222, max=879, avg=585.83, stdev=120.14 00:09:19.804 lat (usec): min=254, max=929, avg=616.82, stdev=122.39 00:09:19.804 clat percentiles (usec): 00:09:19.804 | 1.00th=[ 273], 5.00th=[ 363], 10.00th=[ 437], 20.00th=[ 490], 00:09:19.804 | 30.00th=[ 537], 40.00th=[ 562], 50.00th=[ 594], 60.00th=[ 611], 00:09:19.804 | 70.00th=[ 644], 80.00th=[ 693], 90.00th=[ 734], 95.00th=[ 783], 00:09:19.804 | 99.00th=[ 840], 99.50th=[ 873], 99.90th=[ 881], 99.95th=[ 881], 00:09:19.804 | 99.99th=[ 881] 00:09:19.804 bw ( KiB/s): min= 4096, max= 4096, per=40.30%, avg=4096.00, stdev= 0.00, samples=1 00:09:19.804 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:19.804 lat (usec) : 250=0.45%, 500=16.82%, 750=55.06%, 1000=19.35% 00:09:19.804 lat (msec) : 2=6.40%, 50=1.93% 00:09:19.804 cpu : usr=1.39%, sys=1.88%, ctx=673, majf=0, minf=1 00:09:19.804 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:19.804 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 issued rwts: total=160,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:19.804 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:19.804 job2: (groupid=0, jobs=1): err= 0: pid=4101327: Thu Oct 17 17:30:27 2024 00:09:19.804 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:09:19.804 slat (nsec): min=6676, max=58918, avg=26427.51, stdev=4062.74 00:09:19.804 clat (usec): min=352, max=1394, avg=955.05, stdev=107.72 00:09:19.804 lat (usec): min=360, max=1421, avg=981.48, stdev=108.41 00:09:19.804 clat percentiles (usec): 00:09:19.804 | 1.00th=[ 586], 5.00th=[ 750], 10.00th=[ 840], 20.00th=[ 898], 00:09:19.804 | 30.00th=[ 930], 40.00th=[ 947], 50.00th=[ 971], 60.00th=[ 988], 00:09:19.804 | 70.00th=[ 1004], 80.00th=[ 1029], 90.00th=[ 1057], 95.00th=[ 1090], 00:09:19.804 | 99.00th=[ 1221], 99.50th=[ 1303], 99.90th=[ 1401], 99.95th=[ 1401], 00:09:19.804 | 99.99th=[ 1401] 00:09:19.804 write: IOPS=790, BW=3161KiB/s (3237kB/s)(3164KiB/1001msec); 0 zone resets 00:09:19.804 slat (nsec): min=9065, max=67513, avg=31719.61, stdev=7582.23 00:09:19.804 clat (usec): min=220, max=1072, avg=583.83, stdev=115.77 00:09:19.804 lat (usec): min=230, max=1105, avg=615.55, stdev=118.33 00:09:19.804 clat percentiles (usec): 00:09:19.804 | 1.00th=[ 293], 5.00th=[ 383], 10.00th=[ 441], 20.00th=[ 490], 00:09:19.804 | 30.00th=[ 529], 40.00th=[ 553], 50.00th=[ 586], 60.00th=[ 611], 00:09:19.804 | 70.00th=[ 644], 80.00th=[ 685], 90.00th=[ 725], 95.00th=[ 758], 00:09:19.804 | 99.00th=[ 848], 99.50th=[ 865], 99.90th=[ 1074], 99.95th=[ 1074], 00:09:19.804 | 99.99th=[ 1074] 00:09:19.804 bw ( KiB/s): min= 4096, max= 4096, per=40.30%, avg=4096.00, stdev= 0.00, samples=1 00:09:19.804 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:19.804 lat (usec) : 250=0.31%, 500=13.20%, 750=45.59%, 1000=27.94% 00:09:19.804 lat (msec) : 2=12.97% 00:09:19.804 cpu : usr=3.10%, sys=4.90%, ctx=1303, majf=0, minf=2 00:09:19.804 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:19.804 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 issued rwts: total=512,791,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:19.804 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:19.804 job3: (groupid=0, jobs=1): err= 0: pid=4101328: Thu Oct 17 17:30:27 2024 00:09:19.804 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:09:19.804 slat (nsec): min=25662, max=44889, avg=26807.81, stdev=1915.97 00:09:19.804 clat (usec): min=693, max=1264, avg=990.65, stdev=87.37 00:09:19.804 lat (usec): min=720, max=1291, avg=1017.46, stdev=87.17 00:09:19.804 clat percentiles (usec): 00:09:19.804 | 1.00th=[ 766], 5.00th=[ 824], 10.00th=[ 873], 20.00th=[ 922], 00:09:19.804 | 30.00th=[ 963], 40.00th=[ 988], 50.00th=[ 1004], 60.00th=[ 1020], 00:09:19.804 | 70.00th=[ 1037], 80.00th=[ 1057], 90.00th=[ 1090], 95.00th=[ 1106], 00:09:19.804 | 99.00th=[ 1172], 99.50th=[ 1254], 99.90th=[ 1270], 99.95th=[ 1270], 00:09:19.804 | 99.99th=[ 1270] 00:09:19.804 write: IOPS=786, BW=3145KiB/s (3220kB/s)(3148KiB/1001msec); 0 zone resets 00:09:19.804 slat (nsec): min=9692, max=68627, avg=29808.19, stdev=9448.04 00:09:19.804 clat (usec): min=191, max=3438, avg=565.82, stdev=188.41 00:09:19.804 lat (usec): min=224, max=3471, avg=595.63, stdev=191.39 00:09:19.804 clat percentiles (usec): 00:09:19.804 | 1.00th=[ 273], 5.00th=[ 343], 10.00th=[ 379], 20.00th=[ 453], 00:09:19.804 | 30.00th=[ 494], 40.00th=[ 537], 50.00th=[ 570], 60.00th=[ 603], 00:09:19.804 | 70.00th=[ 627], 80.00th=[ 668], 90.00th=[ 709], 95.00th=[ 742], 00:09:19.804 | 99.00th=[ 807], 99.50th=[ 848], 99.90th=[ 3425], 99.95th=[ 3425], 00:09:19.804 | 99.99th=[ 3425] 00:09:19.804 bw ( KiB/s): min= 4096, max= 4096, per=40.30%, avg=4096.00, stdev= 0.00, samples=1 00:09:19.804 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:09:19.804 lat (usec) : 250=0.38%, 500=19.09%, 750=38.80%, 1000=20.48% 00:09:19.804 lat (msec) : 2=21.09%, 4=0.15% 00:09:19.804 cpu : usr=2.30%, sys=3.50%, ctx=1300, majf=0, minf=1 00:09:19.804 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:19.804 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:19.804 issued rwts: total=512,787,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:19.804 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:19.804 00:09:19.804 Run status group 0 (all jobs): 00:09:19.804 READ: bw=4691KiB/s (4804kB/s), 66.4KiB/s-2046KiB/s (68.0kB/s-2095kB/s), io=4804KiB (4919kB), run=1001-1024msec 00:09:19.804 WRITE: bw=9.93MiB/s (10.4MB/s), 2000KiB/s-3161KiB/s (2048kB/s-3237kB/s), io=10.2MiB (10.7MB), run=1001-1024msec 00:09:19.804 00:09:19.804 Disk stats (read/write): 00:09:19.804 nvme0n1: ios=67/512, merge=0/0, ticks=836/275, in_queue=1111, util=87.37% 00:09:19.804 nvme0n2: ios=205/512, merge=0/0, ticks=779/285, in_queue=1064, util=88.80% 00:09:19.804 nvme0n3: ios=569/546, merge=0/0, ticks=562/222, in_queue=784, util=95.06% 00:09:19.804 nvme0n4: ios=562/530, merge=0/0, ticks=598/289, in_queue=887, util=96.82% 00:09:19.804 17:30:27 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:09:19.804 [global] 00:09:19.804 thread=1 00:09:19.804 invalidate=1 00:09:19.804 rw=write 00:09:19.804 time_based=1 00:09:19.804 runtime=1 00:09:19.804 ioengine=libaio 00:09:19.804 direct=1 00:09:19.804 bs=4096 00:09:19.804 iodepth=128 00:09:19.804 norandommap=0 00:09:19.804 numjobs=1 00:09:19.804 00:09:19.805 verify_dump=1 00:09:19.805 verify_backlog=512 00:09:19.805 verify_state_save=0 00:09:19.805 do_verify=1 00:09:19.805 verify=crc32c-intel 00:09:19.805 [job0] 00:09:19.805 filename=/dev/nvme0n1 00:09:19.805 [job1] 00:09:19.805 filename=/dev/nvme0n2 00:09:19.805 [job2] 00:09:19.805 filename=/dev/nvme0n3 00:09:19.805 [job3] 00:09:19.805 filename=/dev/nvme0n4 00:09:19.805 Could not set queue depth (nvme0n1) 00:09:19.805 Could not set queue depth (nvme0n2) 00:09:19.805 Could not set queue depth (nvme0n3) 00:09:19.805 Could not set queue depth (nvme0n4) 00:09:20.064 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:20.064 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:20.064 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:20.064 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:20.064 fio-3.35 00:09:20.064 Starting 4 threads 00:09:21.448 00:09:21.448 job0: (groupid=0, jobs=1): err= 0: pid=4101854: Thu Oct 17 17:30:29 2024 00:09:21.448 read: IOPS=5059, BW=19.8MiB/s (20.7MB/s)(20.0MiB/1012msec) 00:09:21.448 slat (nsec): min=971, max=12721k, avg=91240.05, stdev=659928.25 00:09:21.448 clat (usec): min=4171, max=44652, avg=10800.15, stdev=5729.43 00:09:21.448 lat (usec): min=4177, max=44661, avg=10891.39, stdev=5784.09 00:09:21.448 clat percentiles (usec): 00:09:21.448 | 1.00th=[ 5145], 5.00th=[ 7701], 10.00th=[ 7898], 20.00th=[ 8029], 00:09:21.448 | 30.00th=[ 8225], 40.00th=[ 8356], 50.00th=[ 8586], 60.00th=[ 8848], 00:09:21.448 | 70.00th=[ 9503], 80.00th=[12780], 90.00th=[17171], 95.00th=[21890], 00:09:21.448 | 99.00th=[39060], 99.50th=[41681], 99.90th=[43779], 99.95th=[44827], 00:09:21.448 | 99.99th=[44827] 00:09:21.448 write: IOPS=5247, BW=20.5MiB/s (21.5MB/s)(20.7MiB/1012msec); 0 zone resets 00:09:21.448 slat (nsec): min=1720, max=10393k, avg=94781.83, stdev=547684.94 00:09:21.448 clat (usec): min=1952, max=45391, avg=13723.37, stdev=9857.49 00:09:21.448 lat (usec): min=1962, max=45394, avg=13818.15, stdev=9916.60 00:09:21.448 clat percentiles (usec): 00:09:21.448 | 1.00th=[ 2900], 5.00th=[ 4621], 10.00th=[ 5735], 20.00th=[ 7046], 00:09:21.448 | 30.00th=[ 7439], 40.00th=[ 8225], 50.00th=[10290], 60.00th=[14222], 00:09:21.448 | 70.00th=[15270], 80.00th=[15533], 90.00th=[28705], 95.00th=[40633], 00:09:21.448 | 99.00th=[44303], 99.50th=[45351], 99.90th=[45351], 99.95th=[45351], 00:09:21.448 | 99.99th=[45351] 00:09:21.448 bw ( KiB/s): min=20672, max=20792, per=20.34%, avg=20732.00, stdev=84.85, samples=2 00:09:21.448 iops : min= 5168, max= 5198, avg=5183.00, stdev=21.21, samples=2 00:09:21.448 lat (msec) : 2=0.10%, 4=1.66%, 10=56.04%, 20=31.54%, 50=10.66% 00:09:21.448 cpu : usr=3.46%, sys=6.33%, ctx=445, majf=0, minf=1 00:09:21.448 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:09:21.448 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:21.448 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:21.448 issued rwts: total=5120,5310,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:21.448 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:21.448 job1: (groupid=0, jobs=1): err= 0: pid=4101856: Thu Oct 17 17:30:29 2024 00:09:21.448 read: IOPS=8815, BW=34.4MiB/s (36.1MB/s)(34.5MiB/1003msec) 00:09:21.448 slat (nsec): min=996, max=7130.6k, avg=58495.85, stdev=432029.63 00:09:21.448 clat (usec): min=1358, max=14437, avg=7596.09, stdev=1569.07 00:09:21.449 lat (usec): min=2379, max=15981, avg=7654.59, stdev=1603.83 00:09:21.449 clat percentiles (usec): 00:09:21.449 | 1.00th=[ 3654], 5.00th=[ 5735], 10.00th=[ 6128], 20.00th=[ 6587], 00:09:21.449 | 30.00th=[ 6849], 40.00th=[ 7111], 50.00th=[ 7242], 60.00th=[ 7504], 00:09:21.449 | 70.00th=[ 7767], 80.00th=[ 8291], 90.00th=[ 9765], 95.00th=[10945], 00:09:21.449 | 99.00th=[12780], 99.50th=[13173], 99.90th=[13304], 99.95th=[13698], 00:09:21.449 | 99.99th=[14484] 00:09:21.449 write: IOPS=9188, BW=35.9MiB/s (37.6MB/s)(36.0MiB/1003msec); 0 zone resets 00:09:21.449 slat (nsec): min=1691, max=5768.2k, avg=47077.12, stdev=323234.71 00:09:21.449 clat (usec): min=1189, max=13648, avg=6507.10, stdev=1424.80 00:09:21.449 lat (usec): min=1205, max=13651, avg=6554.18, stdev=1452.28 00:09:21.449 clat percentiles (usec): 00:09:21.449 | 1.00th=[ 2671], 5.00th=[ 3982], 10.00th=[ 4359], 20.00th=[ 5473], 00:09:21.449 | 30.00th=[ 6259], 40.00th=[ 6652], 50.00th=[ 6783], 60.00th=[ 6980], 00:09:21.449 | 70.00th=[ 7111], 80.00th=[ 7242], 90.00th=[ 7439], 95.00th=[ 9372], 00:09:21.449 | 99.00th=[ 9896], 99.50th=[ 9896], 99.90th=[13042], 99.95th=[13173], 00:09:21.449 | 99.99th=[13698] 00:09:21.449 bw ( KiB/s): min=36864, max=36864, per=36.16%, avg=36864.00, stdev= 0.00, samples=2 00:09:21.449 iops : min= 9216, max= 9216, avg=9216.00, stdev= 0.00, samples=2 00:09:21.449 lat (msec) : 2=0.13%, 4=3.00%, 10=92.33%, 20=4.54% 00:09:21.449 cpu : usr=6.19%, sys=9.68%, ctx=711, majf=0, minf=1 00:09:21.449 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.7% 00:09:21.449 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:21.449 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:21.449 issued rwts: total=8842,9216,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:21.449 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:21.449 job2: (groupid=0, jobs=1): err= 0: pid=4101857: Thu Oct 17 17:30:29 2024 00:09:21.449 read: IOPS=3110, BW=12.2MiB/s (12.7MB/s)(12.3MiB/1010msec) 00:09:21.449 slat (nsec): min=982, max=21989k, avg=112968.05, stdev=858615.42 00:09:21.449 clat (usec): min=4205, max=42704, avg=13218.07, stdev=5407.87 00:09:21.449 lat (usec): min=4211, max=42721, avg=13331.04, stdev=5469.93 00:09:21.449 clat percentiles (usec): 00:09:21.449 | 1.00th=[ 5080], 5.00th=[ 8586], 10.00th=[ 8848], 20.00th=[ 9110], 00:09:21.449 | 30.00th=[ 9503], 40.00th=[ 9765], 50.00th=[10290], 60.00th=[13435], 00:09:21.449 | 70.00th=[15139], 80.00th=[18482], 90.00th=[21103], 95.00th=[22676], 00:09:21.449 | 99.00th=[30016], 99.50th=[30016], 99.90th=[42730], 99.95th=[42730], 00:09:21.449 | 99.99th=[42730] 00:09:21.449 write: IOPS=3548, BW=13.9MiB/s (14.5MB/s)(14.0MiB/1010msec); 0 zone resets 00:09:21.449 slat (nsec): min=1702, max=15569k, avg=175312.67, stdev=853040.07 00:09:21.449 clat (usec): min=2780, max=86456, avg=23743.16, stdev=17543.13 00:09:21.449 lat (usec): min=2790, max=86464, avg=23918.47, stdev=17656.66 00:09:21.449 clat percentiles (usec): 00:09:21.449 | 1.00th=[ 3294], 5.00th=[ 7242], 10.00th=[ 9896], 20.00th=[12780], 00:09:21.449 | 30.00th=[14877], 40.00th=[15270], 50.00th=[15401], 60.00th=[17171], 00:09:21.449 | 70.00th=[26608], 80.00th=[34341], 90.00th=[47449], 95.00th=[65799], 00:09:21.449 | 99.00th=[80217], 99.50th=[84411], 99.90th=[86508], 99.95th=[86508], 00:09:21.449 | 99.99th=[86508] 00:09:21.449 bw ( KiB/s): min=13872, max=14344, per=13.84%, avg=14108.00, stdev=333.75, samples=2 00:09:21.449 iops : min= 3468, max= 3586, avg=3527.00, stdev=83.44, samples=2 00:09:21.449 lat (msec) : 4=0.92%, 10=26.30%, 20=45.91%, 50=21.68%, 100=5.19% 00:09:21.449 cpu : usr=3.47%, sys=2.58%, ctx=439, majf=0, minf=1 00:09:21.449 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.5%, >=64=99.1% 00:09:21.449 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:21.449 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:21.449 issued rwts: total=3142,3584,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:21.449 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:21.449 job3: (groupid=0, jobs=1): err= 0: pid=4101858: Thu Oct 17 17:30:29 2024 00:09:21.449 read: IOPS=7235, BW=28.3MiB/s (29.6MB/s)(28.5MiB/1007msec) 00:09:21.449 slat (nsec): min=996, max=7860.2k, avg=69658.87, stdev=511131.12 00:09:21.449 clat (usec): min=3675, max=16913, avg=9007.68, stdev=2117.29 00:09:21.449 lat (usec): min=3708, max=18833, avg=9077.34, stdev=2150.64 00:09:21.449 clat percentiles (usec): 00:09:21.449 | 1.00th=[ 4490], 5.00th=[ 6521], 10.00th=[ 7046], 20.00th=[ 7635], 00:09:21.449 | 30.00th=[ 7832], 40.00th=[ 8094], 50.00th=[ 8291], 60.00th=[ 8586], 00:09:21.449 | 70.00th=[ 9634], 80.00th=[10552], 90.00th=[12125], 95.00th=[13566], 00:09:21.449 | 99.00th=[15139], 99.50th=[15401], 99.90th=[15926], 99.95th=[16909], 00:09:21.449 | 99.99th=[16909] 00:09:21.449 write: IOPS=7626, BW=29.8MiB/s (31.2MB/s)(30.0MiB/1007msec); 0 zone resets 00:09:21.449 slat (nsec): min=1734, max=31502k, avg=59160.01, stdev=479812.93 00:09:21.449 clat (usec): min=1243, max=32965, avg=8089.12, stdev=3759.29 00:09:21.449 lat (usec): min=1286, max=38603, avg=8148.28, stdev=3777.47 00:09:21.449 clat percentiles (usec): 00:09:21.449 | 1.00th=[ 2933], 5.00th=[ 4080], 10.00th=[ 4948], 20.00th=[ 5735], 00:09:21.449 | 30.00th=[ 7439], 40.00th=[ 7963], 50.00th=[ 8225], 60.00th=[ 8356], 00:09:21.449 | 70.00th=[ 8455], 80.00th=[ 8586], 90.00th=[10028], 95.00th=[11338], 00:09:21.449 | 99.00th=[32900], 99.50th=[32900], 99.90th=[32900], 99.95th=[32900], 00:09:21.449 | 99.99th=[32900] 00:09:21.449 bw ( KiB/s): min=28976, max=32392, per=30.10%, avg=30684.00, stdev=2415.48, samples=2 00:09:21.449 iops : min= 7244, max= 8098, avg=7671.00, stdev=603.87, samples=2 00:09:21.449 lat (msec) : 2=0.03%, 4=2.66%, 10=80.10%, 20=16.36%, 50=0.85% 00:09:21.449 cpu : usr=5.37%, sys=7.06%, ctx=770, majf=0, minf=1 00:09:21.449 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:09:21.449 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:21.449 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:21.449 issued rwts: total=7286,7680,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:21.449 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:21.449 00:09:21.449 Run status group 0 (all jobs): 00:09:21.449 READ: bw=94.1MiB/s (98.7MB/s), 12.2MiB/s-34.4MiB/s (12.7MB/s-36.1MB/s), io=95.3MiB (99.9MB), run=1003-1012msec 00:09:21.449 WRITE: bw=99.5MiB/s (104MB/s), 13.9MiB/s-35.9MiB/s (14.5MB/s-37.6MB/s), io=101MiB (106MB), run=1003-1012msec 00:09:21.449 00:09:21.449 Disk stats (read/write): 00:09:21.449 nvme0n1: ios=4143/4271, merge=0/0, ticks=43171/50401, in_queue=93572, util=86.77% 00:09:21.449 nvme0n2: ios=7148/7168, merge=0/0, ticks=51248/44416, in_queue=95664, util=88.53% 00:09:21.449 nvme0n3: ios=2447/2560, merge=0/0, ticks=33573/63522, in_queue=97095, util=94.85% 00:09:21.449 nvme0n4: ios=5684/6047, merge=0/0, ticks=48716/43590, in_queue=92306, util=98.77% 00:09:21.449 17:30:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:09:21.449 [global] 00:09:21.449 thread=1 00:09:21.449 invalidate=1 00:09:21.449 rw=randwrite 00:09:21.449 time_based=1 00:09:21.449 runtime=1 00:09:21.449 ioengine=libaio 00:09:21.449 direct=1 00:09:21.449 bs=4096 00:09:21.449 iodepth=128 00:09:21.449 norandommap=0 00:09:21.449 numjobs=1 00:09:21.449 00:09:21.449 verify_dump=1 00:09:21.449 verify_backlog=512 00:09:21.449 verify_state_save=0 00:09:21.449 do_verify=1 00:09:21.449 verify=crc32c-intel 00:09:21.449 [job0] 00:09:21.449 filename=/dev/nvme0n1 00:09:21.449 [job1] 00:09:21.449 filename=/dev/nvme0n2 00:09:21.449 [job2] 00:09:21.449 filename=/dev/nvme0n3 00:09:21.449 [job3] 00:09:21.449 filename=/dev/nvme0n4 00:09:21.449 Could not set queue depth (nvme0n1) 00:09:21.449 Could not set queue depth (nvme0n2) 00:09:21.449 Could not set queue depth (nvme0n3) 00:09:21.449 Could not set queue depth (nvme0n4) 00:09:21.708 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:21.708 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:21.708 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:21.708 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:09:21.708 fio-3.35 00:09:21.708 Starting 4 threads 00:09:23.090 00:09:23.090 job0: (groupid=0, jobs=1): err= 0: pid=4102380: Thu Oct 17 17:30:30 2024 00:09:23.090 read: IOPS=6113, BW=23.9MiB/s (25.0MB/s)(24.0MiB/1005msec) 00:09:23.090 slat (nsec): min=939, max=17262k, avg=72606.41, stdev=620285.58 00:09:23.090 clat (usec): min=2869, max=54728, avg=10403.83, stdev=8737.13 00:09:23.090 lat (usec): min=2890, max=56435, avg=10476.44, stdev=8805.66 00:09:23.090 clat percentiles (usec): 00:09:23.090 | 1.00th=[ 3851], 5.00th=[ 5276], 10.00th=[ 5932], 20.00th=[ 6521], 00:09:23.090 | 30.00th=[ 6849], 40.00th=[ 7242], 50.00th=[ 7635], 60.00th=[ 8291], 00:09:23.090 | 70.00th=[ 9241], 80.00th=[10421], 90.00th=[12649], 95.00th=[33817], 00:09:23.090 | 99.00th=[48497], 99.50th=[48497], 99.90th=[48497], 99.95th=[52691], 00:09:23.090 | 99.99th=[54789] 00:09:23.090 write: IOPS=6523, BW=25.5MiB/s (26.7MB/s)(25.6MiB/1005msec); 0 zone resets 00:09:23.090 slat (nsec): min=1587, max=13489k, avg=71956.03, stdev=614951.63 00:09:23.090 clat (usec): min=687, max=39191, avg=9671.40, stdev=6293.99 00:09:23.090 lat (usec): min=710, max=39225, avg=9743.35, stdev=6355.31 00:09:23.090 clat percentiles (usec): 00:09:23.090 | 1.00th=[ 1762], 5.00th=[ 3556], 10.00th=[ 3949], 20.00th=[ 5080], 00:09:23.090 | 30.00th=[ 6521], 40.00th=[ 7046], 50.00th=[ 7701], 60.00th=[ 8029], 00:09:23.090 | 70.00th=[ 9241], 80.00th=[13698], 90.00th=[20055], 95.00th=[24511], 00:09:23.090 | 99.00th=[28443], 99.50th=[30802], 99.90th=[31589], 99.95th=[35390], 00:09:23.090 | 99.99th=[39060] 00:09:23.090 bw ( KiB/s): min=18992, max=32432, per=29.03%, avg=25712.00, stdev=9503.52, samples=2 00:09:23.090 iops : min= 4748, max= 8108, avg=6428.00, stdev=2375.88, samples=2 00:09:23.090 lat (usec) : 750=0.02%, 1000=0.06% 00:09:23.090 lat (msec) : 2=0.64%, 4=5.34%, 10=69.06%, 20=15.71%, 50=9.13% 00:09:23.090 lat (msec) : 100=0.03% 00:09:23.090 cpu : usr=4.88%, sys=7.37%, ctx=354, majf=0, minf=1 00:09:23.090 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.5% 00:09:23.090 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:23.090 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:23.090 issued rwts: total=6144,6556,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:23.090 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:23.090 job1: (groupid=0, jobs=1): err= 0: pid=4102381: Thu Oct 17 17:30:30 2024 00:09:23.090 read: IOPS=3838, BW=15.0MiB/s (15.7MB/s)(15.1MiB/1006msec) 00:09:23.090 slat (nsec): min=1045, max=20080k, avg=143051.45, stdev=1039504.16 00:09:23.090 clat (usec): min=1209, max=51626, avg=18194.39, stdev=11317.14 00:09:23.090 lat (usec): min=3220, max=51652, avg=18337.44, stdev=11407.83 00:09:23.090 clat percentiles (usec): 00:09:23.090 | 1.00th=[ 3326], 5.00th=[ 5342], 10.00th=[ 6980], 20.00th=[ 8356], 00:09:23.090 | 30.00th=[10421], 40.00th=[10814], 50.00th=[11731], 60.00th=[19530], 00:09:23.090 | 70.00th=[25560], 80.00th=[28967], 90.00th=[35914], 95.00th=[38536], 00:09:23.090 | 99.00th=[45351], 99.50th=[45351], 99.90th=[50070], 99.95th=[51119], 00:09:23.090 | 99.99th=[51643] 00:09:23.090 write: IOPS=4071, BW=15.9MiB/s (16.7MB/s)(16.0MiB/1006msec); 0 zone resets 00:09:23.090 slat (nsec): min=1592, max=14172k, avg=104147.09, stdev=837787.72 00:09:23.090 clat (usec): min=1156, max=47702, avg=13889.45, stdev=7541.46 00:09:23.090 lat (usec): min=1164, max=47725, avg=13993.60, stdev=7622.24 00:09:23.090 clat percentiles (usec): 00:09:23.090 | 1.00th=[ 2933], 5.00th=[ 4359], 10.00th=[ 7046], 20.00th=[ 8586], 00:09:23.090 | 30.00th=[ 8979], 40.00th=[ 9372], 50.00th=[10421], 60.00th=[14746], 00:09:23.090 | 70.00th=[17433], 80.00th=[20055], 90.00th=[23725], 95.00th=[27132], 00:09:23.090 | 99.00th=[39584], 99.50th=[39584], 99.90th=[39584], 99.95th=[43254], 00:09:23.090 | 99.99th=[47449] 00:09:23.090 bw ( KiB/s): min=13432, max=19336, per=18.50%, avg=16384.00, stdev=4174.76, samples=2 00:09:23.090 iops : min= 3358, max= 4834, avg=4096.00, stdev=1043.69, samples=2 00:09:23.090 lat (msec) : 2=0.09%, 4=2.39%, 10=35.36%, 20=32.57%, 50=29.54% 00:09:23.090 lat (msec) : 100=0.05% 00:09:23.090 cpu : usr=2.69%, sys=4.68%, ctx=202, majf=0, minf=1 00:09:23.090 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.2% 00:09:23.090 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:23.090 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:23.090 issued rwts: total=3862,4096,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:23.090 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:23.090 job2: (groupid=0, jobs=1): err= 0: pid=4102382: Thu Oct 17 17:30:30 2024 00:09:23.090 read: IOPS=5860, BW=22.9MiB/s (24.0MB/s)(22.9MiB/1002msec) 00:09:23.090 slat (nsec): min=912, max=9992.7k, avg=87695.07, stdev=517911.37 00:09:23.090 clat (usec): min=1328, max=34753, avg=11002.37, stdev=4631.50 00:09:23.090 lat (usec): min=2789, max=35165, avg=11090.06, stdev=4650.75 00:09:23.090 clat percentiles (usec): 00:09:23.090 | 1.00th=[ 3359], 5.00th=[ 7242], 10.00th=[ 7767], 20.00th=[ 8356], 00:09:23.090 | 30.00th=[ 8979], 40.00th=[ 9503], 50.00th=[ 9765], 60.00th=[10159], 00:09:23.090 | 70.00th=[10814], 80.00th=[11731], 90.00th=[16581], 95.00th=[22152], 00:09:23.090 | 99.00th=[31065], 99.50th=[34341], 99.90th=[34866], 99.95th=[34866], 00:09:23.090 | 99.99th=[34866] 00:09:23.090 write: IOPS=6131, BW=24.0MiB/s (25.1MB/s)(24.0MiB/1002msec); 0 zone resets 00:09:23.090 slat (nsec): min=1511, max=10262k, avg=75977.44, stdev=426239.28 00:09:23.090 clat (usec): min=1209, max=26769, avg=10178.12, stdev=4410.50 00:09:23.090 lat (usec): min=1219, max=26773, avg=10254.10, stdev=4438.24 00:09:23.090 clat percentiles (usec): 00:09:23.090 | 1.00th=[ 6194], 5.00th=[ 6587], 10.00th=[ 6783], 20.00th=[ 7111], 00:09:23.090 | 30.00th=[ 8094], 40.00th=[ 8455], 50.00th=[ 9241], 60.00th=[ 9503], 00:09:23.090 | 70.00th=[ 9896], 80.00th=[10421], 90.00th=[19006], 95.00th=[21890], 00:09:23.090 | 99.00th=[24511], 99.50th=[26346], 99.90th=[26870], 99.95th=[26870], 00:09:23.090 | 99.99th=[26870] 00:09:23.090 bw ( KiB/s): min=20480, max=28672, per=27.75%, avg=24576.00, stdev=5792.62, samples=2 00:09:23.090 iops : min= 5120, max= 7168, avg=6144.00, stdev=1448.15, samples=2 00:09:23.090 lat (msec) : 2=0.08%, 4=0.49%, 10=65.41%, 20=26.12%, 50=7.89% 00:09:23.090 cpu : usr=2.60%, sys=3.80%, ctx=669, majf=0, minf=3 00:09:23.090 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.5% 00:09:23.090 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:23.090 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:23.090 issued rwts: total=5872,6144,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:23.090 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:23.090 job3: (groupid=0, jobs=1): err= 0: pid=4102383: Thu Oct 17 17:30:30 2024 00:09:23.090 read: IOPS=5094, BW=19.9MiB/s (20.9MB/s)(20.0MiB/1005msec) 00:09:23.090 slat (nsec): min=1051, max=10128k, avg=84318.40, stdev=613322.06 00:09:23.090 clat (usec): min=5241, max=30257, avg=11100.62, stdev=3307.87 00:09:23.090 lat (usec): min=5248, max=30260, avg=11184.94, stdev=3354.53 00:09:23.090 clat percentiles (usec): 00:09:23.090 | 1.00th=[ 6783], 5.00th=[ 7767], 10.00th=[ 8094], 20.00th=[ 8356], 00:09:23.090 | 30.00th=[ 9110], 40.00th=[ 9503], 50.00th=[10159], 60.00th=[11076], 00:09:23.090 | 70.00th=[11731], 80.00th=[13304], 90.00th=[15270], 95.00th=[18220], 00:09:23.090 | 99.00th=[22676], 99.50th=[25560], 99.90th=[29754], 99.95th=[30278], 00:09:23.090 | 99.99th=[30278] 00:09:23.091 write: IOPS=5451, BW=21.3MiB/s (22.3MB/s)(21.4MiB/1005msec); 0 zone resets 00:09:23.091 slat (nsec): min=1621, max=15545k, avg=97872.93, stdev=643138.20 00:09:23.091 clat (usec): min=2951, max=51427, avg=12846.25, stdev=8630.22 00:09:23.091 lat (usec): min=2959, max=51431, avg=12944.13, stdev=8690.17 00:09:23.091 clat percentiles (usec): 00:09:23.091 | 1.00th=[ 4686], 5.00th=[ 5604], 10.00th=[ 6259], 20.00th=[ 6718], 00:09:23.091 | 30.00th=[ 7373], 40.00th=[ 8160], 50.00th=[ 9241], 60.00th=[11338], 00:09:23.091 | 70.00th=[13042], 80.00th=[18220], 90.00th=[25297], 95.00th=[29754], 00:09:23.091 | 99.00th=[47973], 99.50th=[49546], 99.90th=[51643], 99.95th=[51643], 00:09:23.091 | 99.99th=[51643] 00:09:23.091 bw ( KiB/s): min=20480, max=22336, per=24.17%, avg=21408.00, stdev=1312.39, samples=2 00:09:23.091 iops : min= 5120, max= 5584, avg=5352.00, stdev=328.10, samples=2 00:09:23.091 lat (msec) : 4=0.19%, 10=49.40%, 20=40.16%, 50=10.05%, 100=0.20% 00:09:23.091 cpu : usr=4.28%, sys=5.78%, ctx=337, majf=0, minf=1 00:09:23.091 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:09:23.091 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:23.091 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:09:23.091 issued rwts: total=5120,5479,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:23.091 latency : target=0, window=0, percentile=100.00%, depth=128 00:09:23.091 00:09:23.091 Run status group 0 (all jobs): 00:09:23.091 READ: bw=81.5MiB/s (85.5MB/s), 15.0MiB/s-23.9MiB/s (15.7MB/s-25.0MB/s), io=82.0MiB (86.0MB), run=1002-1006msec 00:09:23.091 WRITE: bw=86.5MiB/s (90.7MB/s), 15.9MiB/s-25.5MiB/s (16.7MB/s-26.7MB/s), io=87.0MiB (91.2MB), run=1002-1006msec 00:09:23.091 00:09:23.091 Disk stats (read/write): 00:09:23.091 nvme0n1: ios=5871/6144, merge=0/0, ticks=34624/32507, in_queue=67131, util=85.07% 00:09:23.091 nvme0n2: ios=3385/3584, merge=0/0, ticks=29532/22368, in_queue=51900, util=88.80% 00:09:23.091 nvme0n3: ios=4665/5024, merge=0/0, ticks=16642/19386, in_queue=36028, util=95.06% 00:09:23.091 nvme0n4: ios=4157/4443, merge=0/0, ticks=42978/54248, in_queue=97226, util=94.69% 00:09:23.091 17:30:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:09:23.091 17:30:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=4102717 00:09:23.091 17:30:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:09:23.091 17:30:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:09:23.091 [global] 00:09:23.091 thread=1 00:09:23.091 invalidate=1 00:09:23.091 rw=read 00:09:23.091 time_based=1 00:09:23.091 runtime=10 00:09:23.091 ioengine=libaio 00:09:23.091 direct=1 00:09:23.091 bs=4096 00:09:23.091 iodepth=1 00:09:23.091 norandommap=1 00:09:23.091 numjobs=1 00:09:23.091 00:09:23.091 [job0] 00:09:23.091 filename=/dev/nvme0n1 00:09:23.091 [job1] 00:09:23.091 filename=/dev/nvme0n2 00:09:23.091 [job2] 00:09:23.091 filename=/dev/nvme0n3 00:09:23.091 [job3] 00:09:23.091 filename=/dev/nvme0n4 00:09:23.091 Could not set queue depth (nvme0n1) 00:09:23.091 Could not set queue depth (nvme0n2) 00:09:23.091 Could not set queue depth (nvme0n3) 00:09:23.091 Could not set queue depth (nvme0n4) 00:09:23.659 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:23.659 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:23.659 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:23.659 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:09:23.659 fio-3.35 00:09:23.659 Starting 4 threads 00:09:26.202 17:30:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:09:26.202 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=6725632, buflen=4096 00:09:26.202 fio: pid=4102911, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:09:26.202 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:09:26.462 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=10756096, buflen=4096 00:09:26.462 fio: pid=4102910, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:09:26.462 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:09:26.463 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:09:26.723 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:09:26.723 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:09:26.723 fio: io_u error on file /dev/nvme0n1: Operation not supported: read offset=11313152, buflen=4096 00:09:26.723 fio: pid=4102904, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:09:26.723 fio: io_u error on file /dev/nvme0n2: Operation not supported: read offset=16605184, buflen=4096 00:09:26.723 fio: pid=4102905, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:09:26.723 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:09:26.723 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:09:26.984 00:09:26.984 job0: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=4102904: Thu Oct 17 17:30:34 2024 00:09:26.984 read: IOPS=935, BW=3739KiB/s (3828kB/s)(10.8MiB/2955msec) 00:09:26.984 slat (usec): min=8, max=34527, avg=47.62, stdev=749.07 00:09:26.984 clat (usec): min=491, max=6249, avg=1007.74, stdev=145.75 00:09:26.984 lat (usec): min=516, max=35603, avg=1050.67, stdev=723.97 00:09:26.984 clat percentiles (usec): 00:09:26.984 | 1.00th=[ 709], 5.00th=[ 816], 10.00th=[ 865], 20.00th=[ 930], 00:09:26.984 | 30.00th=[ 971], 40.00th=[ 996], 50.00th=[ 1012], 60.00th=[ 1037], 00:09:26.984 | 70.00th=[ 1057], 80.00th=[ 1090], 90.00th=[ 1139], 95.00th=[ 1172], 00:09:26.984 | 99.00th=[ 1237], 99.50th=[ 1270], 99.90th=[ 1336], 99.95th=[ 1352], 00:09:26.984 | 99.99th=[ 6259] 00:09:26.984 bw ( KiB/s): min= 3664, max= 3912, per=27.16%, avg=3846.40, stdev=105.28, samples=5 00:09:26.984 iops : min= 916, max= 978, avg=961.60, stdev=26.32, samples=5 00:09:26.984 lat (usec) : 500=0.04%, 750=1.77%, 1000=40.79% 00:09:26.984 lat (msec) : 2=57.33%, 10=0.04% 00:09:26.984 cpu : usr=0.78%, sys=3.01%, ctx=2766, majf=0, minf=1 00:09:26.984 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:26.984 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 issued rwts: total=2763,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:26.984 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:26.984 job1: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=4102905: Thu Oct 17 17:30:34 2024 00:09:26.984 read: IOPS=1295, BW=5179KiB/s (5303kB/s)(15.8MiB/3131msec) 00:09:26.984 slat (usec): min=6, max=22190, avg=39.63, stdev=512.00 00:09:26.984 clat (usec): min=164, max=2487, avg=720.16, stdev=152.83 00:09:26.984 lat (usec): min=171, max=23120, avg=759.80, stdev=535.73 00:09:26.984 clat percentiles (usec): 00:09:26.984 | 1.00th=[ 289], 5.00th=[ 453], 10.00th=[ 519], 20.00th=[ 594], 00:09:26.984 | 30.00th=[ 635], 40.00th=[ 701], 50.00th=[ 742], 60.00th=[ 775], 00:09:26.984 | 70.00th=[ 824], 80.00th=[ 857], 90.00th=[ 898], 95.00th=[ 922], 00:09:26.984 | 99.00th=[ 971], 99.50th=[ 996], 99.90th=[ 1057], 99.95th=[ 1090], 00:09:26.984 | 99.99th=[ 2474] 00:09:26.984 bw ( KiB/s): min= 5000, max= 5631, per=37.16%, avg=5262.00, stdev=256.40, samples=6 00:09:26.984 iops : min= 1250, max= 1407, avg=1315.33, stdev=63.91, samples=6 00:09:26.984 lat (usec) : 250=0.44%, 500=7.99%, 750=42.98%, 1000=48.14% 00:09:26.984 lat (msec) : 2=0.39%, 4=0.02% 00:09:26.984 cpu : usr=1.18%, sys=3.77%, ctx=4060, majf=0, minf=2 00:09:26.984 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:26.984 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 issued rwts: total=4055,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:26.984 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:26.984 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=4102910: Thu Oct 17 17:30:34 2024 00:09:26.984 read: IOPS=953, BW=3811KiB/s (3903kB/s)(10.3MiB/2756msec) 00:09:26.984 slat (usec): min=7, max=21708, avg=39.56, stdev=513.63 00:09:26.984 clat (usec): min=499, max=1316, avg=993.45, stdev=99.52 00:09:26.984 lat (usec): min=525, max=22749, avg=1033.02, stdev=524.05 00:09:26.984 clat percentiles (usec): 00:09:26.984 | 1.00th=[ 734], 5.00th=[ 799], 10.00th=[ 857], 20.00th=[ 914], 00:09:26.984 | 30.00th=[ 955], 40.00th=[ 988], 50.00th=[ 1012], 60.00th=[ 1037], 00:09:26.984 | 70.00th=[ 1057], 80.00th=[ 1074], 90.00th=[ 1106], 95.00th=[ 1123], 00:09:26.984 | 99.00th=[ 1188], 99.50th=[ 1205], 99.90th=[ 1237], 99.95th=[ 1287], 00:09:26.984 | 99.99th=[ 1319] 00:09:26.984 bw ( KiB/s): min= 3832, max= 4008, per=27.52%, avg=3897.60, stdev=73.84, samples=5 00:09:26.984 iops : min= 958, max= 1002, avg=974.40, stdev=18.46, samples=5 00:09:26.984 lat (usec) : 500=0.04%, 750=1.75%, 1000=42.44% 00:09:26.984 lat (msec) : 2=55.73% 00:09:26.984 cpu : usr=1.13%, sys=2.76%, ctx=2629, majf=0, minf=2 00:09:26.984 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:26.984 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 issued rwts: total=2627,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:26.984 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:26.984 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=4102911: Thu Oct 17 17:30:34 2024 00:09:26.984 read: IOPS=638, BW=2551KiB/s (2612kB/s)(6568KiB/2575msec) 00:09:26.984 slat (nsec): min=24452, max=57999, avg=25696.66, stdev=2723.03 00:09:26.984 clat (usec): min=562, max=42719, avg=1519.48, stdev=4573.82 00:09:26.984 lat (usec): min=587, max=42744, avg=1545.17, stdev=4573.79 00:09:26.984 clat percentiles (usec): 00:09:26.984 | 1.00th=[ 734], 5.00th=[ 807], 10.00th=[ 840], 20.00th=[ 914], 00:09:26.984 | 30.00th=[ 955], 40.00th=[ 988], 50.00th=[ 1012], 60.00th=[ 1037], 00:09:26.984 | 70.00th=[ 1057], 80.00th=[ 1090], 90.00th=[ 1139], 95.00th=[ 1188], 00:09:26.984 | 99.00th=[41157], 99.50th=[42206], 99.90th=[42206], 99.95th=[42730], 00:09:26.984 | 99.99th=[42730] 00:09:26.984 bw ( KiB/s): min= 103, max= 3944, per=18.06%, avg=2558.20, stdev=1853.39, samples=5 00:09:26.984 iops : min= 25, max= 986, avg=639.40, stdev=463.60, samples=5 00:09:26.984 lat (usec) : 750=1.83%, 1000=43.82% 00:09:26.984 lat (msec) : 2=53.01%, 50=1.28% 00:09:26.984 cpu : usr=0.85%, sys=1.75%, ctx=1643, majf=0, minf=1 00:09:26.984 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:09:26.984 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:09:26.984 issued rwts: total=1643,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:09:26.984 latency : target=0, window=0, percentile=100.00%, depth=1 00:09:26.984 00:09:26.984 Run status group 0 (all jobs): 00:09:26.984 READ: bw=13.8MiB/s (14.5MB/s), 2551KiB/s-5179KiB/s (2612kB/s-5303kB/s), io=43.3MiB (45.4MB), run=2575-3131msec 00:09:26.984 00:09:26.984 Disk stats (read/write): 00:09:26.984 nvme0n1: ios=2641/0, merge=0/0, ticks=2612/0, in_queue=2612, util=91.69% 00:09:26.984 nvme0n2: ios=4000/0, merge=0/0, ticks=2775/0, in_queue=2775, util=93.59% 00:09:26.984 nvme0n3: ios=2475/0, merge=0/0, ticks=2402/0, in_queue=2402, util=95.64% 00:09:26.984 nvme0n4: ios=1635/0, merge=0/0, ticks=2416/0, in_queue=2416, util=96.36% 00:09:26.984 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:09:26.984 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:09:27.245 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:09:27.245 17:30:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:09:27.245 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:09:27.506 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:09:27.506 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:09:27.506 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:09:27.766 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # wait 4102717 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:09:27.767 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1219 -- # local i=0 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1231 -- # return 0 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:09:27.767 nvmf hotplug test: fio failed as expected 00:09:27.767 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:09:28.027 rmmod nvme_tcp 00:09:28.027 rmmod nvme_fabrics 00:09:28.027 rmmod nvme_keyring 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@515 -- # '[' -n 4098901 ']' 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@516 -- # killprocess 4098901 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@950 -- # '[' -z 4098901 ']' 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@954 -- # kill -0 4098901 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@955 -- # uname 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:28.027 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4098901 00:09:28.287 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:28.287 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:28.287 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4098901' 00:09:28.287 killing process with pid 4098901 00:09:28.287 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@969 -- # kill 4098901 00:09:28.287 17:30:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@974 -- # wait 4098901 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-restore 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-save 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:28.287 17:30:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:09:30.831 00:09:30.831 real 0m29.557s 00:09:30.831 user 2m28.552s 00:09:30.831 sys 0m10.155s 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:09:30.831 ************************************ 00:09:30.831 END TEST nvmf_fio_target 00:09:30.831 ************************************ 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:09:30.831 ************************************ 00:09:30.831 START TEST nvmf_bdevio 00:09:30.831 ************************************ 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:09:30.831 * Looking for test storage... 00:09:30.831 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1691 -- # lcov --version 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:09:30.831 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.831 --rc genhtml_branch_coverage=1 00:09:30.831 --rc genhtml_function_coverage=1 00:09:30.831 --rc genhtml_legend=1 00:09:30.831 --rc geninfo_all_blocks=1 00:09:30.831 --rc geninfo_unexecuted_blocks=1 00:09:30.831 00:09:30.831 ' 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:09:30.831 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.831 --rc genhtml_branch_coverage=1 00:09:30.831 --rc genhtml_function_coverage=1 00:09:30.831 --rc genhtml_legend=1 00:09:30.831 --rc geninfo_all_blocks=1 00:09:30.831 --rc geninfo_unexecuted_blocks=1 00:09:30.831 00:09:30.831 ' 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:09:30.831 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.831 --rc genhtml_branch_coverage=1 00:09:30.831 --rc genhtml_function_coverage=1 00:09:30.831 --rc genhtml_legend=1 00:09:30.831 --rc geninfo_all_blocks=1 00:09:30.831 --rc geninfo_unexecuted_blocks=1 00:09:30.831 00:09:30.831 ' 00:09:30.831 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:09:30.831 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.831 --rc genhtml_branch_coverage=1 00:09:30.831 --rc genhtml_function_coverage=1 00:09:30.832 --rc genhtml_legend=1 00:09:30.832 --rc geninfo_all_blocks=1 00:09:30.832 --rc geninfo_unexecuted_blocks=1 00:09:30.832 00:09:30.832 ' 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:30.832 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@474 -- # prepare_net_devs 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@436 -- # local -g is_hw=no 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@438 -- # remove_spdk_ns 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:09:30.832 17:30:38 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:09:38.970 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:09:38.971 Found 0000:31:00.0 (0x8086 - 0x159b) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:09:38.971 Found 0000:31:00.1 (0x8086 - 0x159b) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:09:38.971 Found net devices under 0000:31:00.0: cvl_0_0 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:09:38.971 Found net devices under 0000:31:00.1: cvl_0_1 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # is_hw=yes 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:38.971 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:38.971 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.756 ms 00:09:38.971 00:09:38.971 --- 10.0.0.2 ping statistics --- 00:09:38.971 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:38.971 rtt min/avg/max/mdev = 0.756/0.756/0.756/0.000 ms 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:38.971 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:38.971 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.293 ms 00:09:38.971 00:09:38.971 --- 10.0.0.1 ping statistics --- 00:09:38.971 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:38.971 rtt min/avg/max/mdev = 0.293/0.293/0.293/0.000 ms 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@448 -- # return 0 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:09:38.971 17:30:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@724 -- # xtrace_disable 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@507 -- # nvmfpid=4108040 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@508 -- # waitforlisten 4108040 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x78 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@831 -- # '[' -z 4108040 ']' 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:38.971 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:38.972 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:38.972 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:38.972 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:38.972 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:38.972 [2024-10-17 17:30:46.089801] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:09:38.972 [2024-10-17 17:30:46.089887] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:38.972 [2024-10-17 17:30:46.181483] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:38.972 [2024-10-17 17:30:46.232480] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:38.972 [2024-10-17 17:30:46.232533] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:38.972 [2024-10-17 17:30:46.232542] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:38.972 [2024-10-17 17:30:46.232549] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:38.972 [2024-10-17 17:30:46.232556] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:38.972 [2024-10-17 17:30:46.234646] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:09:38.972 [2024-10-17 17:30:46.234814] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:09:38.972 [2024-10-17 17:30:46.235099] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:09:38.972 [2024-10-17 17:30:46.235103] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@864 -- # return 0 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@730 -- # xtrace_disable 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:39.233 [2024-10-17 17:30:46.974223] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.233 17:30:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:39.233 Malloc0 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:39.233 [2024-10-17 17:30:47.040682] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@558 -- # config=() 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@558 -- # local subsystem config 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:09:39.233 { 00:09:39.233 "params": { 00:09:39.233 "name": "Nvme$subsystem", 00:09:39.233 "trtype": "$TEST_TRANSPORT", 00:09:39.233 "traddr": "$NVMF_FIRST_TARGET_IP", 00:09:39.233 "adrfam": "ipv4", 00:09:39.233 "trsvcid": "$NVMF_PORT", 00:09:39.233 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:09:39.233 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:09:39.233 "hdgst": ${hdgst:-false}, 00:09:39.233 "ddgst": ${ddgst:-false} 00:09:39.233 }, 00:09:39.233 "method": "bdev_nvme_attach_controller" 00:09:39.233 } 00:09:39.233 EOF 00:09:39.233 )") 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@580 -- # cat 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@582 -- # jq . 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@583 -- # IFS=, 00:09:39.233 17:30:47 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:09:39.233 "params": { 00:09:39.233 "name": "Nvme1", 00:09:39.233 "trtype": "tcp", 00:09:39.233 "traddr": "10.0.0.2", 00:09:39.233 "adrfam": "ipv4", 00:09:39.233 "trsvcid": "4420", 00:09:39.233 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:09:39.233 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:09:39.233 "hdgst": false, 00:09:39.233 "ddgst": false 00:09:39.233 }, 00:09:39.233 "method": "bdev_nvme_attach_controller" 00:09:39.233 }' 00:09:39.233 [2024-10-17 17:30:47.097757] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:09:39.233 [2024-10-17 17:30:47.097827] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4108339 ] 00:09:39.498 [2024-10-17 17:30:47.182321] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:09:39.498 [2024-10-17 17:30:47.238732] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:39.498 [2024-10-17 17:30:47.238896] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:39.498 [2024-10-17 17:30:47.238992] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:39.760 I/O targets: 00:09:39.760 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:09:39.760 00:09:39.760 00:09:39.760 CUnit - A unit testing framework for C - Version 2.1-3 00:09:39.760 http://cunit.sourceforge.net/ 00:09:39.760 00:09:39.760 00:09:39.760 Suite: bdevio tests on: Nvme1n1 00:09:39.760 Test: blockdev write read block ...passed 00:09:39.760 Test: blockdev write zeroes read block ...passed 00:09:39.760 Test: blockdev write zeroes read no split ...passed 00:09:39.760 Test: blockdev write zeroes read split ...passed 00:09:40.021 Test: blockdev write zeroes read split partial ...passed 00:09:40.021 Test: blockdev reset ...[2024-10-17 17:30:47.729528] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:09:40.021 [2024-10-17 17:30:47.729624] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xbd1eb0 (9): Bad file descriptor 00:09:40.021 [2024-10-17 17:30:47.743063] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:09:40.021 passed 00:09:40.021 Test: blockdev write read 8 blocks ...passed 00:09:40.021 Test: blockdev write read size > 128k ...passed 00:09:40.021 Test: blockdev write read invalid size ...passed 00:09:40.021 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:09:40.021 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:09:40.021 Test: blockdev write read max offset ...passed 00:09:40.283 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:09:40.283 Test: blockdev writev readv 8 blocks ...passed 00:09:40.283 Test: blockdev writev readv 30 x 1block ...passed 00:09:40.283 Test: blockdev writev readv block ...passed 00:09:40.283 Test: blockdev writev readv size > 128k ...passed 00:09:40.283 Test: blockdev writev readv size > 128k in two iovs ...passed 00:09:40.283 Test: blockdev comparev and writev ...[2024-10-17 17:30:48.004178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.004229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.004246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.004255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.004704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.004718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.004733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.004742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.005157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.005169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.005183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.005193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.005632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.005643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.005657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:09:40.283 [2024-10-17 17:30:48.005665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:09:40.283 passed 00:09:40.283 Test: blockdev nvme passthru rw ...passed 00:09:40.283 Test: blockdev nvme passthru vendor specific ...[2024-10-17 17:30:48.089988] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:09:40.283 [2024-10-17 17:30:48.090006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.090230] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:09:40.283 [2024-10-17 17:30:48.090241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.090514] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:09:40.283 [2024-10-17 17:30:48.090527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:09:40.283 [2024-10-17 17:30:48.090790] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:09:40.283 [2024-10-17 17:30:48.090801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:09:40.283 passed 00:09:40.283 Test: blockdev nvme admin passthru ...passed 00:09:40.283 Test: blockdev copy ...passed 00:09:40.283 00:09:40.283 Run Summary: Type Total Ran Passed Failed Inactive 00:09:40.283 suites 1 1 n/a 0 0 00:09:40.283 tests 23 23 23 0 0 00:09:40.283 asserts 152 152 152 0 n/a 00:09:40.283 00:09:40.283 Elapsed time = 1.266 seconds 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@514 -- # nvmfcleanup 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:09:40.544 rmmod nvme_tcp 00:09:40.544 rmmod nvme_fabrics 00:09:40.544 rmmod nvme_keyring 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@515 -- # '[' -n 4108040 ']' 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@516 -- # killprocess 4108040 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@950 -- # '[' -z 4108040 ']' 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@954 -- # kill -0 4108040 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@955 -- # uname 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4108040 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@956 -- # process_name=reactor_3 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@960 -- # '[' reactor_3 = sudo ']' 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4108040' 00:09:40.544 killing process with pid 4108040 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@969 -- # kill 4108040 00:09:40.544 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@974 -- # wait 4108040 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-save 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-restore 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:40.806 17:30:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:42.717 17:30:50 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:09:42.717 00:09:42.717 real 0m12.380s 00:09:42.717 user 0m13.946s 00:09:42.717 sys 0m6.251s 00:09:42.717 17:30:50 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:42.717 17:30:50 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:09:42.717 ************************************ 00:09:42.717 END TEST nvmf_bdevio 00:09:42.717 ************************************ 00:09:42.978 17:30:50 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:09:42.978 00:09:42.978 real 5m6.100s 00:09:42.978 user 11m39.260s 00:09:42.978 sys 1m50.627s 00:09:42.978 17:30:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:42.978 17:30:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:09:42.978 ************************************ 00:09:42.978 END TEST nvmf_target_core 00:09:42.978 ************************************ 00:09:42.978 17:30:50 nvmf_tcp -- nvmf/nvmf.sh@15 -- # run_test nvmf_target_extra /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:09:42.978 17:30:50 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:09:42.978 17:30:50 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:42.978 17:30:50 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:09:42.978 ************************************ 00:09:42.978 START TEST nvmf_target_extra 00:09:42.978 ************************************ 00:09:42.978 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:09:42.978 * Looking for test storage... 00:09:42.978 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:09:42.978 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:09:42.978 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1691 -- # lcov --version 00:09:42.978 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # IFS=.-: 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # read -ra ver1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # IFS=.-: 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # read -ra ver2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@338 -- # local 'op=<' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@340 -- # ver1_l=2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@341 -- # ver2_l=1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@344 -- # case "$op" in 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@345 -- # : 1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # decimal 1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # ver1[v]=1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # decimal 2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # ver2[v]=2 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # return 0 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:09:43.239 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.239 --rc genhtml_branch_coverage=1 00:09:43.239 --rc genhtml_function_coverage=1 00:09:43.239 --rc genhtml_legend=1 00:09:43.239 --rc geninfo_all_blocks=1 00:09:43.239 --rc geninfo_unexecuted_blocks=1 00:09:43.239 00:09:43.239 ' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:09:43.239 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.239 --rc genhtml_branch_coverage=1 00:09:43.239 --rc genhtml_function_coverage=1 00:09:43.239 --rc genhtml_legend=1 00:09:43.239 --rc geninfo_all_blocks=1 00:09:43.239 --rc geninfo_unexecuted_blocks=1 00:09:43.239 00:09:43.239 ' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:09:43.239 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.239 --rc genhtml_branch_coverage=1 00:09:43.239 --rc genhtml_function_coverage=1 00:09:43.239 --rc genhtml_legend=1 00:09:43.239 --rc geninfo_all_blocks=1 00:09:43.239 --rc geninfo_unexecuted_blocks=1 00:09:43.239 00:09:43.239 ' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:09:43.239 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.239 --rc genhtml_branch_coverage=1 00:09:43.239 --rc genhtml_function_coverage=1 00:09:43.239 --rc genhtml_legend=1 00:09:43.239 --rc geninfo_all_blocks=1 00:09:43.239 --rc geninfo_unexecuted_blocks=1 00:09:43.239 00:09:43.239 ' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # uname -s 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@15 -- # shopt -s extglob 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- paths/export.sh@5 -- # export PATH 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@51 -- # : 0 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:43.239 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:43.240 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@13 -- # TEST_ARGS=("$@") 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@15 -- # [[ 0 -eq 0 ]] 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@16 -- # run_test nvmf_example /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:43.240 17:30:50 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:09:43.240 ************************************ 00:09:43.240 START TEST nvmf_example 00:09:43.240 ************************************ 00:09:43.240 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:09:43.240 * Looking for test storage... 00:09:43.240 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:43.240 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:09:43.240 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1691 -- # lcov --version 00:09:43.240 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # IFS=.-: 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # read -ra ver1 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # IFS=.-: 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # read -ra ver2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@338 -- # local 'op=<' 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@340 -- # ver1_l=2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@341 -- # ver2_l=1 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@344 -- # case "$op" in 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@345 -- # : 1 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # decimal 1 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=1 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 1 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # ver1[v]=1 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # decimal 2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # ver2[v]=2 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:43.501 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # return 0 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:09:43.502 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.502 --rc genhtml_branch_coverage=1 00:09:43.502 --rc genhtml_function_coverage=1 00:09:43.502 --rc genhtml_legend=1 00:09:43.502 --rc geninfo_all_blocks=1 00:09:43.502 --rc geninfo_unexecuted_blocks=1 00:09:43.502 00:09:43.502 ' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:09:43.502 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.502 --rc genhtml_branch_coverage=1 00:09:43.502 --rc genhtml_function_coverage=1 00:09:43.502 --rc genhtml_legend=1 00:09:43.502 --rc geninfo_all_blocks=1 00:09:43.502 --rc geninfo_unexecuted_blocks=1 00:09:43.502 00:09:43.502 ' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:09:43.502 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.502 --rc genhtml_branch_coverage=1 00:09:43.502 --rc genhtml_function_coverage=1 00:09:43.502 --rc genhtml_legend=1 00:09:43.502 --rc geninfo_all_blocks=1 00:09:43.502 --rc geninfo_unexecuted_blocks=1 00:09:43.502 00:09:43.502 ' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:09:43.502 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:43.502 --rc genhtml_branch_coverage=1 00:09:43.502 --rc genhtml_function_coverage=1 00:09:43.502 --rc genhtml_legend=1 00:09:43.502 --rc geninfo_all_blocks=1 00:09:43.502 --rc geninfo_unexecuted_blocks=1 00:09:43.502 00:09:43.502 ' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # uname -s 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@15 -- # shopt -s extglob 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@5 -- # export PATH 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@51 -- # : 0 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:43.502 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@11 -- # NVMF_EXAMPLE=("$SPDK_EXAMPLE_DIR/nvmf") 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@13 -- # MALLOC_BDEV_SIZE=64 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@24 -- # build_nvmf_example_args 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@17 -- # '[' 0 -eq 1 ']' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@20 -- # NVMF_EXAMPLE+=(-i "$NVMF_APP_SHM_ID" -g 10000) 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@21 -- # NVMF_EXAMPLE+=("${NO_HUGE[@]}") 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@40 -- # timing_enter nvmf_example_test 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@724 -- # xtrace_disable 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@41 -- # nvmftestinit 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@474 -- # prepare_net_devs 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@436 -- # local -g is_hw=no 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@438 -- # remove_spdk_ns 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@309 -- # xtrace_disable 00:09:43.502 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # pci_devs=() 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # net_devs=() 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # e810=() 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # local -ga e810 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # x722=() 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # local -ga x722 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # mlx=() 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # local -ga mlx 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:09:51.644 Found 0000:31:00.0 (0x8086 - 0x159b) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:09:51.644 Found 0000:31:00.1 (0x8086 - 0x159b) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:09:51.644 Found net devices under 0000:31:00.0: cvl_0_0 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:09:51.644 Found net devices under 0000:31:00.1: cvl_0_1 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # is_hw=yes 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:51.644 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:51.645 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:51.645 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.713 ms 00:09:51.645 00:09:51.645 --- 10.0.0.2 ping statistics --- 00:09:51.645 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:51.645 rtt min/avg/max/mdev = 0.713/0.713/0.713/0.000 ms 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:51.645 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:51.645 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.320 ms 00:09:51.645 00:09:51.645 --- 10.0.0.1 ping statistics --- 00:09:51.645 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:51.645 rtt min/avg/max/mdev = 0.320/0.320/0.320/0.000 ms 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@448 -- # return 0 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@42 -- # nvmfexamplestart '-m 0xF' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@27 -- # timing_enter start_nvmf_example 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@724 -- # xtrace_disable 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@29 -- # '[' tcp == tcp ']' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@30 -- # NVMF_EXAMPLE=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_EXAMPLE[@]}") 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@34 -- # nvmfpid=4113093 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@35 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/nvmf -i 0 -g 10000 -m 0xF 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@36 -- # waitforlisten 4113093 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@831 -- # '[' -z 4113093 ']' 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:51.645 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:51.645 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@864 -- # return 0 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@37 -- # timing_exit start_nvmf_example 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@730 -- # xtrace_disable 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # rpc_cmd bdev_malloc_create 64 512 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # malloc_bdevs='Malloc0 ' 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@49 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@52 -- # for malloc_bdev in $malloc_bdevs 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@59 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:09:52.216 17:30:59 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:10:02.219 Initializing NVMe Controllers 00:10:02.219 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:02.219 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:10:02.219 Initialization complete. Launching workers. 00:10:02.219 ======================================================== 00:10:02.219 Latency(us) 00:10:02.219 Device Information : IOPS MiB/s Average min max 00:10:02.219 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 18842.71 73.60 3396.22 635.64 19352.61 00:10:02.219 ======================================================== 00:10:02.219 Total : 18842.71 73.60 3396.22 635.64 19352.61 00:10:02.219 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@65 -- # trap - SIGINT SIGTERM EXIT 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@66 -- # nvmftestfini 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@514 -- # nvmfcleanup 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@121 -- # sync 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@124 -- # set +e 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:02.480 rmmod nvme_tcp 00:10:02.480 rmmod nvme_fabrics 00:10:02.480 rmmod nvme_keyring 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@128 -- # set -e 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@129 -- # return 0 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@515 -- # '[' -n 4113093 ']' 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@516 -- # killprocess 4113093 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@950 -- # '[' -z 4113093 ']' 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@954 -- # kill -0 4113093 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@955 -- # uname 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4113093 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@956 -- # process_name=nvmf 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@960 -- # '[' nvmf = sudo ']' 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4113093' 00:10:02.480 killing process with pid 4113093 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@969 -- # kill 4113093 00:10:02.480 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@974 -- # wait 4113093 00:10:02.739 nvmf threads initialize successfully 00:10:02.739 bdev subsystem init successfully 00:10:02.739 created a nvmf target service 00:10:02.739 create targets's poll groups done 00:10:02.739 all subsystems of target started 00:10:02.739 nvmf target is running 00:10:02.739 all subsystems of target stopped 00:10:02.739 destroy targets's poll groups done 00:10:02.739 destroyed the nvmf target service 00:10:02.739 bdev subsystem finish successfully 00:10:02.739 nvmf threads destroy successfully 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@297 -- # iptr 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@789 -- # iptables-save 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@789 -- # iptables-restore 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:02.739 17:31:10 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:04.663 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:04.663 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@67 -- # timing_exit nvmf_example_test 00:10:04.663 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@730 -- # xtrace_disable 00:10:04.663 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:04.663 00:10:04.663 real 0m21.517s 00:10:04.663 user 0m46.471s 00:10:04.663 sys 0m7.079s 00:10:04.663 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:04.663 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:10:04.663 ************************************ 00:10:04.663 END TEST nvmf_example 00:10:04.663 ************************************ 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@17 -- # run_test nvmf_filesystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:10:04.926 ************************************ 00:10:04.926 START TEST nvmf_filesystem 00:10:04.926 ************************************ 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:10:04.926 * Looking for test storage... 00:10:04.926 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1691 -- # lcov --version 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:10:04.926 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:04.926 --rc genhtml_branch_coverage=1 00:10:04.926 --rc genhtml_function_coverage=1 00:10:04.926 --rc genhtml_legend=1 00:10:04.926 --rc geninfo_all_blocks=1 00:10:04.926 --rc geninfo_unexecuted_blocks=1 00:10:04.926 00:10:04.926 ' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:10:04.926 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:04.926 --rc genhtml_branch_coverage=1 00:10:04.926 --rc genhtml_function_coverage=1 00:10:04.926 --rc genhtml_legend=1 00:10:04.926 --rc geninfo_all_blocks=1 00:10:04.926 --rc geninfo_unexecuted_blocks=1 00:10:04.926 00:10:04.926 ' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:10:04.926 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:04.926 --rc genhtml_branch_coverage=1 00:10:04.926 --rc genhtml_function_coverage=1 00:10:04.926 --rc genhtml_legend=1 00:10:04.926 --rc geninfo_all_blocks=1 00:10:04.926 --rc geninfo_unexecuted_blocks=1 00:10:04.926 00:10:04.926 ' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:10:04.926 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:04.926 --rc genhtml_branch_coverage=1 00:10:04.926 --rc genhtml_function_coverage=1 00:10:04.926 --rc genhtml_legend=1 00:10:04.926 --rc geninfo_all_blocks=1 00:10:04.926 --rc geninfo_unexecuted_blocks=1 00:10:04.926 00:10:04.926 ' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@34 -- # set -e 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@36 -- # shopt -s extglob 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@37 -- # shopt -s inherit_errexit 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@39 -- # '[' -z /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output ']' 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@44 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh ]] 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@17 -- # CONFIG_PGO_CAPTURE=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@18 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@19 -- # CONFIG_ENV=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@20 -- # CONFIG_LTO=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@21 -- # CONFIG_ISCSI_INITIATOR=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@22 -- # CONFIG_CET=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@23 -- # CONFIG_VBDEV_COMPRESS_MLX5=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@24 -- # CONFIG_OCF_PATH= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@25 -- # CONFIG_RDMA_SET_TOS=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@26 -- # CONFIG_AIO_FSDEV=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@27 -- # CONFIG_HAVE_ARC4RANDOM=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@28 -- # CONFIG_HAVE_LIBARCHIVE=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@29 -- # CONFIG_UBLK=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@30 -- # CONFIG_ISAL_CRYPTO=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@31 -- # CONFIG_OPENSSL_PATH= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@32 -- # CONFIG_OCF=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@33 -- # CONFIG_FUSE=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@34 -- # CONFIG_VTUNE_DIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@35 -- # CONFIG_FUZZER_LIB= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@36 -- # CONFIG_FUZZER=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@37 -- # CONFIG_FSDEV=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@38 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@39 -- # CONFIG_CRYPTO=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@40 -- # CONFIG_PGO_USE=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@41 -- # CONFIG_VHOST=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@42 -- # CONFIG_DAOS=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@43 -- # CONFIG_DPDK_INC_DIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@44 -- # CONFIG_DAOS_DIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@45 -- # CONFIG_UNIT_TESTS=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@46 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@47 -- # CONFIG_VIRTIO=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@48 -- # CONFIG_DPDK_UADK=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@49 -- # CONFIG_COVERAGE=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@50 -- # CONFIG_RDMA=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@51 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIM=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@52 -- # CONFIG_HAVE_LZ4=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@53 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@54 -- # CONFIG_URING_PATH= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@55 -- # CONFIG_XNVME=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@56 -- # CONFIG_VFIO_USER=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@57 -- # CONFIG_ARCH=native 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@58 -- # CONFIG_HAVE_EVP_MAC=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@59 -- # CONFIG_URING_ZNS=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@60 -- # CONFIG_WERROR=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@61 -- # CONFIG_HAVE_LIBBSD=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@62 -- # CONFIG_UBSAN=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@63 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@64 -- # CONFIG_IPSEC_MB_DIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@65 -- # CONFIG_GOLANG=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@66 -- # CONFIG_ISAL=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@67 -- # CONFIG_IDXD_KERNEL=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@68 -- # CONFIG_DPDK_LIB_DIR= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@69 -- # CONFIG_RDMA_PROV=verbs 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@70 -- # CONFIG_APPS=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@71 -- # CONFIG_SHARED=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@72 -- # CONFIG_HAVE_KEYUTILS=y 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@73 -- # CONFIG_FC_PATH= 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@74 -- # CONFIG_DPDK_PKG_CONFIG=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@75 -- # CONFIG_FC=n 00:10:04.926 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@76 -- # CONFIG_AVAHI=n 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@77 -- # CONFIG_FIO_PLUGIN=y 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@78 -- # CONFIG_RAID5F=n 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@79 -- # CONFIG_EXAMPLES=y 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@80 -- # CONFIG_TESTS=y 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@81 -- # CONFIG_CRYPTO_MLX5=n 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@82 -- # CONFIG_MAX_LCORES=128 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@83 -- # CONFIG_IPSEC_MB=n 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@84 -- # CONFIG_PGO_DIR= 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@85 -- # CONFIG_DEBUG=y 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@86 -- # CONFIG_DPDK_COMPRESSDEV=n 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@87 -- # CONFIG_CROSS_PREFIX= 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@88 -- # CONFIG_COPY_FILE_RANGE=y 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@89 -- # CONFIG_URING=n 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:10:04.927 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/config.h ]] 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:10:05.190 #define SPDK_CONFIG_H 00:10:05.190 #define SPDK_CONFIG_AIO_FSDEV 1 00:10:05.190 #define SPDK_CONFIG_APPS 1 00:10:05.190 #define SPDK_CONFIG_ARCH native 00:10:05.190 #undef SPDK_CONFIG_ASAN 00:10:05.190 #undef SPDK_CONFIG_AVAHI 00:10:05.190 #undef SPDK_CONFIG_CET 00:10:05.190 #define SPDK_CONFIG_COPY_FILE_RANGE 1 00:10:05.190 #define SPDK_CONFIG_COVERAGE 1 00:10:05.190 #define SPDK_CONFIG_CROSS_PREFIX 00:10:05.190 #undef SPDK_CONFIG_CRYPTO 00:10:05.190 #undef SPDK_CONFIG_CRYPTO_MLX5 00:10:05.190 #undef SPDK_CONFIG_CUSTOMOCF 00:10:05.190 #undef SPDK_CONFIG_DAOS 00:10:05.190 #define SPDK_CONFIG_DAOS_DIR 00:10:05.190 #define SPDK_CONFIG_DEBUG 1 00:10:05.190 #undef SPDK_CONFIG_DPDK_COMPRESSDEV 00:10:05.190 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:10:05.190 #define SPDK_CONFIG_DPDK_INC_DIR 00:10:05.190 #define SPDK_CONFIG_DPDK_LIB_DIR 00:10:05.190 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:10:05.190 #undef SPDK_CONFIG_DPDK_UADK 00:10:05.190 #define SPDK_CONFIG_ENV /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:10:05.190 #define SPDK_CONFIG_EXAMPLES 1 00:10:05.190 #undef SPDK_CONFIG_FC 00:10:05.190 #define SPDK_CONFIG_FC_PATH 00:10:05.190 #define SPDK_CONFIG_FIO_PLUGIN 1 00:10:05.190 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:10:05.190 #define SPDK_CONFIG_FSDEV 1 00:10:05.190 #undef SPDK_CONFIG_FUSE 00:10:05.190 #undef SPDK_CONFIG_FUZZER 00:10:05.190 #define SPDK_CONFIG_FUZZER_LIB 00:10:05.190 #undef SPDK_CONFIG_GOLANG 00:10:05.190 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:10:05.190 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:10:05.190 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:10:05.190 #define SPDK_CONFIG_HAVE_KEYUTILS 1 00:10:05.190 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:10:05.190 #undef SPDK_CONFIG_HAVE_LIBBSD 00:10:05.190 #undef SPDK_CONFIG_HAVE_LZ4 00:10:05.190 #define SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIM 1 00:10:05.190 #undef SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC 00:10:05.190 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:10:05.190 #define SPDK_CONFIG_IDXD 1 00:10:05.190 #define SPDK_CONFIG_IDXD_KERNEL 1 00:10:05.190 #undef SPDK_CONFIG_IPSEC_MB 00:10:05.190 #define SPDK_CONFIG_IPSEC_MB_DIR 00:10:05.190 #define SPDK_CONFIG_ISAL 1 00:10:05.190 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:10:05.190 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:10:05.190 #define SPDK_CONFIG_LIBDIR 00:10:05.190 #undef SPDK_CONFIG_LTO 00:10:05.190 #define SPDK_CONFIG_MAX_LCORES 128 00:10:05.190 #define SPDK_CONFIG_NVME_CUSE 1 00:10:05.190 #undef SPDK_CONFIG_OCF 00:10:05.190 #define SPDK_CONFIG_OCF_PATH 00:10:05.190 #define SPDK_CONFIG_OPENSSL_PATH 00:10:05.190 #undef SPDK_CONFIG_PGO_CAPTURE 00:10:05.190 #define SPDK_CONFIG_PGO_DIR 00:10:05.190 #undef SPDK_CONFIG_PGO_USE 00:10:05.190 #define SPDK_CONFIG_PREFIX /usr/local 00:10:05.190 #undef SPDK_CONFIG_RAID5F 00:10:05.190 #undef SPDK_CONFIG_RBD 00:10:05.190 #define SPDK_CONFIG_RDMA 1 00:10:05.190 #define SPDK_CONFIG_RDMA_PROV verbs 00:10:05.190 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:10:05.190 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:10:05.190 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:10:05.190 #define SPDK_CONFIG_SHARED 1 00:10:05.190 #undef SPDK_CONFIG_SMA 00:10:05.190 #define SPDK_CONFIG_TESTS 1 00:10:05.190 #undef SPDK_CONFIG_TSAN 00:10:05.190 #define SPDK_CONFIG_UBLK 1 00:10:05.190 #define SPDK_CONFIG_UBSAN 1 00:10:05.190 #undef SPDK_CONFIG_UNIT_TESTS 00:10:05.190 #undef SPDK_CONFIG_URING 00:10:05.190 #define SPDK_CONFIG_URING_PATH 00:10:05.190 #undef SPDK_CONFIG_URING_ZNS 00:10:05.190 #undef SPDK_CONFIG_USDT 00:10:05.190 #undef SPDK_CONFIG_VBDEV_COMPRESS 00:10:05.190 #undef SPDK_CONFIG_VBDEV_COMPRESS_MLX5 00:10:05.190 #define SPDK_CONFIG_VFIO_USER 1 00:10:05.190 #define SPDK_CONFIG_VFIO_USER_DIR 00:10:05.190 #define SPDK_CONFIG_VHOST 1 00:10:05.190 #define SPDK_CONFIG_VIRTIO 1 00:10:05.190 #undef SPDK_CONFIG_VTUNE 00:10:05.190 #define SPDK_CONFIG_VTUNE_DIR 00:10:05.190 #define SPDK_CONFIG_WERROR 1 00:10:05.190 #define SPDK_CONFIG_WPDK_DIR 00:10:05.190 #undef SPDK_CONFIG_XNVME 00:10:05.190 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:05.190 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@56 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/../../../ 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@64 -- # TEST_TAG=N/A 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@65 -- # TEST_TAG_FILE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.run_test_name 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@67 -- # PM_OUTPUTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # uname -s 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # PM_OS=Linux 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # MONITOR_RESOURCES_SUDO=() 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # declare -A MONITOR_RESOURCES_SUDO 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@71 -- # MONITOR_RESOURCES_SUDO["collect-bmc-pm"]=1 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@72 -- # MONITOR_RESOURCES_SUDO["collect-cpu-load"]=0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@73 -- # MONITOR_RESOURCES_SUDO["collect-cpu-temp"]=0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@74 -- # MONITOR_RESOURCES_SUDO["collect-vmstat"]=0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[0]= 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[1]='sudo -E' 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@78 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@79 -- # [[ Linux == FreeBSD ]] 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ Linux == Linux ]] 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ............................... != QEMU ]] 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ! -e /.dockerenv ]] 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@84 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@85 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@88 -- # [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power ]] 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@58 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@59 -- # export RUN_NIGHTLY 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@62 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@63 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@64 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@65 -- # export SPDK_RUN_VALGRIND 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@66 -- # : 1 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@67 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@68 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@69 -- # export SPDK_TEST_UNITTEST 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@70 -- # : 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@71 -- # export SPDK_TEST_AUTOBUILD 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@72 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@73 -- # export SPDK_TEST_RELEASE_BUILD 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@74 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@75 -- # export SPDK_TEST_ISAL 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@76 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@77 -- # export SPDK_TEST_ISCSI 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@78 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@79 -- # export SPDK_TEST_ISCSI_INITIATOR 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@80 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@81 -- # export SPDK_TEST_NVME 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@82 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@83 -- # export SPDK_TEST_NVME_PMR 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@84 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@85 -- # export SPDK_TEST_NVME_BP 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@86 -- # : 1 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@87 -- # export SPDK_TEST_NVME_CLI 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@88 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@89 -- # export SPDK_TEST_NVME_CUSE 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@90 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@91 -- # export SPDK_TEST_NVME_FDP 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@92 -- # : 1 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@93 -- # export SPDK_TEST_NVMF 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@94 -- # : 1 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@95 -- # export SPDK_TEST_VFIOUSER 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@96 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@97 -- # export SPDK_TEST_VFIOUSER_QEMU 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@98 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@99 -- # export SPDK_TEST_FUZZER 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@100 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@101 -- # export SPDK_TEST_FUZZER_SHORT 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@102 -- # : tcp 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@103 -- # export SPDK_TEST_NVMF_TRANSPORT 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@104 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@105 -- # export SPDK_TEST_RBD 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@106 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@107 -- # export SPDK_TEST_VHOST 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@108 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@109 -- # export SPDK_TEST_BLOCKDEV 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@110 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@111 -- # export SPDK_TEST_RAID 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@112 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@113 -- # export SPDK_TEST_IOAT 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@114 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@115 -- # export SPDK_TEST_BLOBFS 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@116 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@117 -- # export SPDK_TEST_VHOST_INIT 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@118 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@119 -- # export SPDK_TEST_LVOL 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@120 -- # : 0 00:10:05.191 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@121 -- # export SPDK_TEST_VBDEV_COMPRESS 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@122 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@123 -- # export SPDK_RUN_ASAN 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@124 -- # : 1 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@125 -- # export SPDK_RUN_UBSAN 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@126 -- # : 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@127 -- # export SPDK_RUN_EXTERNAL_DPDK 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@128 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@129 -- # export SPDK_RUN_NON_ROOT 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@130 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@131 -- # export SPDK_TEST_CRYPTO 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@132 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@133 -- # export SPDK_TEST_FTL 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@134 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@135 -- # export SPDK_TEST_OCF 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@136 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@137 -- # export SPDK_TEST_VMD 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@138 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@139 -- # export SPDK_TEST_OPAL 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@140 -- # : 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@141 -- # export SPDK_TEST_NATIVE_DPDK 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@142 -- # : true 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@143 -- # export SPDK_AUTOTEST_X 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@144 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@145 -- # export SPDK_TEST_URING 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@146 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@147 -- # export SPDK_TEST_USDT 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@148 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@149 -- # export SPDK_TEST_USE_IGB_UIO 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@150 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@151 -- # export SPDK_TEST_SCHEDULER 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@152 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@153 -- # export SPDK_TEST_SCANBUILD 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@154 -- # : e810 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@155 -- # export SPDK_TEST_NVMF_NICS 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@156 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@157 -- # export SPDK_TEST_SMA 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@158 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@159 -- # export SPDK_TEST_DAOS 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@160 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@161 -- # export SPDK_TEST_XNVME 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@162 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@163 -- # export SPDK_TEST_ACCEL 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@164 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@165 -- # export SPDK_TEST_ACCEL_DSA 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@166 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@167 -- # export SPDK_TEST_ACCEL_IAA 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@169 -- # : 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@170 -- # export SPDK_TEST_FUZZER_TARGET 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@171 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@172 -- # export SPDK_TEST_NVMF_MDNS 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@173 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@174 -- # export SPDK_JSONRPC_GO_CLIENT 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@175 -- # : 0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@176 -- # export SPDK_TEST_SETUP 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@179 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@179 -- # SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@180 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@180 -- # DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@185 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@185 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@189 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@189 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@193 -- # export PYTHONDONTWRITEBYTECODE=1 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@193 -- # PYTHONDONTWRITEBYTECODE=1 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@197 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@197 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@198 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@198 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@202 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@203 -- # rm -rf /var/tmp/asan_suppression_file 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@204 -- # cat 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@240 -- # echo leak:libfuse3.so 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@242 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@242 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@246 -- # '[' -z /var/spdk/dependencies ']' 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@249 -- # export DEPENDENCY_DIR 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@253 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@253 -- # SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:05.192 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@254 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@254 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@257 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@257 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@258 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@258 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # export AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@263 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@263 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@265 -- # _LCOV_MAIN=0 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@266 -- # _LCOV_LLVM=1 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@267 -- # _LCOV= 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@268 -- # [[ '' == *clang* ]] 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@268 -- # [[ 0 -eq 1 ]] 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@270 -- # _lcov_opt[_LCOV_LLVM]='--gcov-tool /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/fuzz/llvm/llvm-gcov.sh' 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@271 -- # _lcov_opt[_LCOV_MAIN]= 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@273 -- # lcov_opt= 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@276 -- # '[' 0 -eq 0 ']' 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@277 -- # export valgrind= 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@277 -- # valgrind= 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@283 -- # uname -s 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@283 -- # '[' Linux = Linux ']' 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@284 -- # HUGEMEM=4096 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # export CLEAR_HUGE=yes 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # CLEAR_HUGE=yes 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@287 -- # MAKE=make 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@288 -- # MAKEFLAGS=-j144 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@304 -- # export HUGEMEM=4096 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@304 -- # HUGEMEM=4096 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@306 -- # NO_HUGE=() 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@307 -- # TEST_MODE= 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@308 -- # for i in "$@" 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@309 -- # case "$i" in 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@314 -- # TEST_TRANSPORT=tcp 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@329 -- # [[ -z 4115863 ]] 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@329 -- # kill -0 4115863 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1676 -- # set_test_storage 2147483648 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@339 -- # [[ -v testdir ]] 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@341 -- # local requested_size=2147483648 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@342 -- # local mount target_dir 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@344 -- # local -A mounts fss sizes avails uses 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@345 -- # local source fs size avail mount use 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@347 -- # local storage_fallback storage_candidates 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@349 -- # mktemp -udt spdk.XXXXXX 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@349 -- # storage_fallback=/tmp/spdk.qthWAq 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@354 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@356 -- # [[ -n '' ]] 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@361 -- # [[ -n '' ]] 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@366 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target /tmp/spdk.qthWAq/tests/target /tmp/spdk.qthWAq 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@369 -- # requested_size=2214592512 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@338 -- # df -T 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@338 -- # grep -v Filesystem 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=spdk_devtmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=devtmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=67108864 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=67108864 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=0 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=spdk_root 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=overlay 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=126797983744 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=134640029696 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=7842045952 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=67309981696 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=67320012800 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=10031104 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=26904645632 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=26928009216 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=23363584 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=efivarfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=efivarfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=387072 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=507904 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=116736 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=67319361536 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=67320016896 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=655360 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=13463990272 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=13464002560 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=12288 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@377 -- # printf '* Looking for test storage...\n' 00:10:05.193 * Looking for test storage... 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@379 -- # local target_space new_size 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@380 -- # for target_dir in "${storage_candidates[@]}" 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@383 -- # df /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@383 -- # awk '$1 !~ /Filesystem/{print $6}' 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@383 -- # mount=/ 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # target_space=126797983744 00:10:05.193 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@386 -- # (( target_space == 0 || target_space < requested_size )) 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@389 -- # (( target_space >= requested_size )) 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # [[ overlay == tmpfs ]] 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # [[ overlay == ramfs ]] 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # [[ / == / ]] 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@392 -- # new_size=10056638464 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # (( new_size * 100 / sizes[/] > 95 )) 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@398 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@398 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@399 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:05.194 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@400 -- # return 0 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1678 -- # set -o errtrace 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1679 -- # shopt -s extdebug 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1680 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1682 -- # PS4=' \t ${test_domain:-} -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1683 -- # true 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1685 -- # xtrace_fd 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -n 15 ]] 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/15 ]] 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@27 -- # exec 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@29 -- # exec 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@31 -- # xtrace_restore 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@18 -- # set -x 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1691 -- # lcov --version 00:10:05.194 17:31:12 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:10:05.194 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.194 --rc genhtml_branch_coverage=1 00:10:05.194 --rc genhtml_function_coverage=1 00:10:05.194 --rc genhtml_legend=1 00:10:05.194 --rc geninfo_all_blocks=1 00:10:05.194 --rc geninfo_unexecuted_blocks=1 00:10:05.194 00:10:05.194 ' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:10:05.194 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.194 --rc genhtml_branch_coverage=1 00:10:05.194 --rc genhtml_function_coverage=1 00:10:05.194 --rc genhtml_legend=1 00:10:05.194 --rc geninfo_all_blocks=1 00:10:05.194 --rc geninfo_unexecuted_blocks=1 00:10:05.194 00:10:05.194 ' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:10:05.194 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.194 --rc genhtml_branch_coverage=1 00:10:05.194 --rc genhtml_function_coverage=1 00:10:05.194 --rc genhtml_legend=1 00:10:05.194 --rc geninfo_all_blocks=1 00:10:05.194 --rc geninfo_unexecuted_blocks=1 00:10:05.194 00:10:05.194 ' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:10:05.194 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.194 --rc genhtml_branch_coverage=1 00:10:05.194 --rc genhtml_function_coverage=1 00:10:05.194 --rc genhtml_legend=1 00:10:05.194 --rc geninfo_all_blocks=1 00:10:05.194 --rc geninfo_unexecuted_blocks=1 00:10:05.194 00:10:05.194 ' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # uname -s 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:05.194 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:05.195 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:05.195 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:10:05.456 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:05.456 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:05.456 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:05.456 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.456 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@51 -- # : 0 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:05.457 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@12 -- # MALLOC_BDEV_SIZE=512 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@15 -- # nvmftestinit 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@474 -- # prepare_net_devs 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@436 -- # local -g is_hw=no 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@438 -- # remove_spdk_ns 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@309 -- # xtrace_disable 00:10:05.457 17:31:13 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # pci_devs=() 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # net_devs=() 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # e810=() 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # local -ga e810 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # x722=() 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # local -ga x722 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # mlx=() 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # local -ga mlx 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:10:13.710 Found 0000:31:00.0 (0x8086 - 0x159b) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:10:13.710 Found 0000:31:00.1 (0x8086 - 0x159b) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:13.710 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:10:13.711 Found net devices under 0000:31:00.0: cvl_0_0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:10:13.711 Found net devices under 0000:31:00.1: cvl_0_1 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # is_hw=yes 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:13.711 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:13.711 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.675 ms 00:10:13.711 00:10:13.711 --- 10.0.0.2 ping statistics --- 00:10:13.711 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:13.711 rtt min/avg/max/mdev = 0.675/0.675/0.675/0.000 ms 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:13.711 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:13.711 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.276 ms 00:10:13.711 00:10:13.711 --- 10.0.0.1 ping statistics --- 00:10:13.711 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:13.711 rtt min/avg/max/mdev = 0.276/0.276/0.276/0.000 ms 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@448 -- # return 0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@105 -- # run_test nvmf_filesystem_no_in_capsule nvmf_filesystem_part 0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:10:13.711 ************************************ 00:10:13.711 START TEST nvmf_filesystem_no_in_capsule 00:10:13.711 ************************************ 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1125 -- # nvmf_filesystem_part 0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@47 -- # in_capsule=0 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@724 -- # xtrace_disable 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@507 -- # nvmfpid=4119561 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@508 -- # waitforlisten 4119561 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@831 -- # '[' -z 4119561 ']' 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:13.711 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:13.711 17:31:20 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:13.711 [2024-10-17 17:31:20.904911] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:10:13.711 [2024-10-17 17:31:20.904975] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:13.711 [2024-10-17 17:31:20.997678] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:13.711 [2024-10-17 17:31:21.052807] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:13.711 [2024-10-17 17:31:21.052856] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:13.711 [2024-10-17 17:31:21.052865] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:13.711 [2024-10-17 17:31:21.052872] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:13.711 [2024-10-17 17:31:21.052879] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:13.711 [2024-10-17 17:31:21.055004] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:13.711 [2024-10-17 17:31:21.055228] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:13.711 [2024-10-17 17:31:21.055386] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:10:13.711 [2024-10-17 17:31:21.055389] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@864 -- # return 0 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@730 -- # xtrace_disable 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:13.972 [2024-10-17 17:31:21.779907] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.972 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:14.233 Malloc1 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:14.233 [2024-10-17 17:31:21.927517] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1378 -- # local bdev_name=Malloc1 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1379 -- # local bdev_info 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1380 -- # local bs 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1381 -- # local nb 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.233 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # bdev_info='[ 00:10:14.233 { 00:10:14.233 "name": "Malloc1", 00:10:14.233 "aliases": [ 00:10:14.233 "696091bc-6830-4c37-a389-a38c6d3160c8" 00:10:14.233 ], 00:10:14.233 "product_name": "Malloc disk", 00:10:14.233 "block_size": 512, 00:10:14.233 "num_blocks": 1048576, 00:10:14.233 "uuid": "696091bc-6830-4c37-a389-a38c6d3160c8", 00:10:14.233 "assigned_rate_limits": { 00:10:14.233 "rw_ios_per_sec": 0, 00:10:14.233 "rw_mbytes_per_sec": 0, 00:10:14.233 "r_mbytes_per_sec": 0, 00:10:14.233 "w_mbytes_per_sec": 0 00:10:14.233 }, 00:10:14.233 "claimed": true, 00:10:14.233 "claim_type": "exclusive_write", 00:10:14.233 "zoned": false, 00:10:14.233 "supported_io_types": { 00:10:14.233 "read": true, 00:10:14.233 "write": true, 00:10:14.233 "unmap": true, 00:10:14.233 "flush": true, 00:10:14.233 "reset": true, 00:10:14.233 "nvme_admin": false, 00:10:14.233 "nvme_io": false, 00:10:14.233 "nvme_io_md": false, 00:10:14.233 "write_zeroes": true, 00:10:14.233 "zcopy": true, 00:10:14.233 "get_zone_info": false, 00:10:14.233 "zone_management": false, 00:10:14.233 "zone_append": false, 00:10:14.233 "compare": false, 00:10:14.233 "compare_and_write": false, 00:10:14.233 "abort": true, 00:10:14.233 "seek_hole": false, 00:10:14.233 "seek_data": false, 00:10:14.233 "copy": true, 00:10:14.233 "nvme_iov_md": false 00:10:14.233 }, 00:10:14.233 "memory_domains": [ 00:10:14.233 { 00:10:14.233 "dma_device_id": "system", 00:10:14.233 "dma_device_type": 1 00:10:14.233 }, 00:10:14.233 { 00:10:14.233 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.233 "dma_device_type": 2 00:10:14.233 } 00:10:14.234 ], 00:10:14.234 "driver_specific": {} 00:10:14.234 } 00:10:14.234 ]' 00:10:14.234 17:31:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # jq '.[] .block_size' 00:10:14.234 17:31:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # bs=512 00:10:14.234 17:31:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # jq '.[] .num_blocks' 00:10:14.234 17:31:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # nb=1048576 00:10:14.234 17:31:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1387 -- # bdev_size=512 00:10:14.234 17:31:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1388 -- # echo 512 00:10:14.234 17:31:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:10:14.234 17:31:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:10:16.148 17:31:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:10:16.148 17:31:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1198 -- # local i=0 00:10:16.148 17:31:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:10:16.148 17:31:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:10:16.148 17:31:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1205 -- # sleep 2 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1208 -- # return 0 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:10:18.059 17:31:25 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:10:19.000 17:31:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@76 -- # '[' 0 -eq 0 ']' 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@77 -- # run_test filesystem_ext4 nvmf_filesystem_create ext4 nvme0n1 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:19.938 ************************************ 00:10:19.938 START TEST filesystem_ext4 00:10:19.938 ************************************ 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create ext4 nvme0n1 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@926 -- # local fstype=ext4 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@928 -- # local i=0 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@929 -- # local force 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@931 -- # '[' ext4 = ext4 ']' 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@932 -- # force=-F 00:10:19.938 17:31:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@937 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:10:19.938 mke2fs 1.47.0 (5-Feb-2023) 00:10:19.938 Discarding device blocks: 0/522240 done 00:10:19.938 Creating filesystem with 522240 1k blocks and 130560 inodes 00:10:19.938 Filesystem UUID: 43e21529-793a-4061-b87e-6eb024b89c11 00:10:19.938 Superblock backups stored on blocks: 00:10:19.939 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:10:19.939 00:10:19.939 Allocating group tables: 0/64 done 00:10:19.939 Writing inode tables: 0/64 done 00:10:20.199 Creating journal (8192 blocks): done 00:10:22.411 Writing superblocks and filesystem accounting information: 0/6426/64 done 00:10:22.411 00:10:22.411 17:31:30 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@945 -- # return 0 00:10:22.411 17:31:30 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@25 -- # sync 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@27 -- # sync 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@29 -- # i=0 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@37 -- # kill -0 4119561 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:28.990 00:10:28.990 real 0m8.178s 00:10:28.990 user 0m0.020s 00:10:28.990 sys 0m0.091s 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@10 -- # set +x 00:10:28.990 ************************************ 00:10:28.990 END TEST filesystem_ext4 00:10:28.990 ************************************ 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@78 -- # run_test filesystem_btrfs nvmf_filesystem_create btrfs nvme0n1 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:28.990 ************************************ 00:10:28.990 START TEST filesystem_btrfs 00:10:28.990 ************************************ 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create btrfs nvme0n1 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@926 -- # local fstype=btrfs 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@928 -- # local i=0 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@929 -- # local force 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@931 -- # '[' btrfs = ext4 ']' 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@934 -- # force=-f 00:10:28.990 17:31:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@937 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:10:28.990 btrfs-progs v6.8.1 00:10:28.990 See https://btrfs.readthedocs.io for more information. 00:10:28.990 00:10:28.990 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:10:28.990 NOTE: several default settings have changed in version 5.15, please make sure 00:10:28.990 this does not affect your deployments: 00:10:28.990 - DUP for metadata (-m dup) 00:10:28.990 - enabled no-holes (-O no-holes) 00:10:28.990 - enabled free-space-tree (-R free-space-tree) 00:10:28.990 00:10:28.990 Label: (null) 00:10:28.990 UUID: 618132d9-61b1-4975-b396-f60249a749e2 00:10:28.990 Node size: 16384 00:10:28.990 Sector size: 4096 (CPU page size: 4096) 00:10:28.990 Filesystem size: 510.00MiB 00:10:28.990 Block group profiles: 00:10:28.990 Data: single 8.00MiB 00:10:28.990 Metadata: DUP 32.00MiB 00:10:28.990 System: DUP 8.00MiB 00:10:28.990 SSD detected: yes 00:10:28.990 Zoned device: no 00:10:28.990 Features: extref, skinny-metadata, no-holes, free-space-tree 00:10:28.990 Checksum: crc32c 00:10:28.990 Number of devices: 1 00:10:28.990 Devices: 00:10:28.990 ID SIZE PATH 00:10:28.990 1 510.00MiB /dev/nvme0n1p1 00:10:28.991 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@945 -- # return 0 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@25 -- # sync 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@27 -- # sync 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@29 -- # i=0 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@37 -- # kill -0 4119561 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:28.991 00:10:28.991 real 0m0.810s 00:10:28.991 user 0m0.033s 00:10:28.991 sys 0m0.114s 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@10 -- # set +x 00:10:28.991 ************************************ 00:10:28.991 END TEST filesystem_btrfs 00:10:28.991 ************************************ 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@79 -- # run_test filesystem_xfs nvmf_filesystem_create xfs nvme0n1 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:28.991 ************************************ 00:10:28.991 START TEST filesystem_xfs 00:10:28.991 ************************************ 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create xfs nvme0n1 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@926 -- # local fstype=xfs 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@928 -- # local i=0 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@929 -- # local force 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@931 -- # '[' xfs = ext4 ']' 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@934 -- # force=-f 00:10:28.991 17:31:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@937 -- # mkfs.xfs -f /dev/nvme0n1p1 00:10:28.991 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:10:28.991 = sectsz=512 attr=2, projid32bit=1 00:10:28.991 = crc=1 finobt=1, sparse=1, rmapbt=0 00:10:28.991 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:10:28.991 data = bsize=4096 blocks=130560, imaxpct=25 00:10:28.991 = sunit=0 swidth=0 blks 00:10:28.991 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:10:28.991 log =internal log bsize=4096 blocks=16384, version=2 00:10:28.991 = sectsz=512 sunit=0 blks, lazy-count=1 00:10:28.991 realtime =none extsz=4096 blocks=0, rtextents=0 00:10:29.929 Discarding blocks...Done. 00:10:29.929 17:31:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@945 -- # return 0 00:10:29.929 17:31:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:32.469 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:32.469 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@25 -- # sync 00:10:32.469 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@27 -- # sync 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@29 -- # i=0 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@37 -- # kill -0 4119561 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:32.470 00:10:32.470 real 0m3.376s 00:10:32.470 user 0m0.021s 00:10:32.470 sys 0m0.084s 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@10 -- # set +x 00:10:32.470 ************************************ 00:10:32.470 END TEST filesystem_xfs 00:10:32.470 ************************************ 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:10:32.470 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@93 -- # sync 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:33.040 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1219 -- # local i=0 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1231 -- # return 0 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.040 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@101 -- # killprocess 4119561 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@950 -- # '[' -z 4119561 ']' 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@954 -- # kill -0 4119561 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@955 -- # uname 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4119561 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4119561' 00:10:33.041 killing process with pid 4119561 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@969 -- # kill 4119561 00:10:33.041 17:31:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@974 -- # wait 4119561 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:10:33.301 00:10:33.301 real 0m20.262s 00:10:33.301 user 1m20.073s 00:10:33.301 sys 0m1.485s 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:33.301 ************************************ 00:10:33.301 END TEST nvmf_filesystem_no_in_capsule 00:10:33.301 ************************************ 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@106 -- # run_test nvmf_filesystem_in_capsule nvmf_filesystem_part 4096 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:10:33.301 ************************************ 00:10:33.301 START TEST nvmf_filesystem_in_capsule 00:10:33.301 ************************************ 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1125 -- # nvmf_filesystem_part 4096 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@47 -- # in_capsule=4096 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@724 -- # xtrace_disable 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@507 -- # nvmfpid=4123804 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@508 -- # waitforlisten 4123804 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@831 -- # '[' -z 4123804 ']' 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:33.301 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:33.301 17:31:41 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:33.561 [2024-10-17 17:31:41.246787] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:10:33.561 [2024-10-17 17:31:41.246840] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:33.561 [2024-10-17 17:31:41.333017] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:33.561 [2024-10-17 17:31:41.368817] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:33.561 [2024-10-17 17:31:41.368847] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:33.561 [2024-10-17 17:31:41.368853] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:33.561 [2024-10-17 17:31:41.368858] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:33.561 [2024-10-17 17:31:41.368862] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:33.561 [2024-10-17 17:31:41.370182] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:33.561 [2024-10-17 17:31:41.370333] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:33.561 [2024-10-17 17:31:41.370467] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:33.561 [2024-10-17 17:31:41.370469] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@864 -- # return 0 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@730 -- # xtrace_disable 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 4096 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:34.502 [2024-10-17 17:31:42.100516] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:34.502 Malloc1 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:34.502 [2024-10-17 17:31:42.224784] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1378 -- # local bdev_name=Malloc1 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1379 -- # local bdev_info 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1380 -- # local bs 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1381 -- # local nb 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.502 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # bdev_info='[ 00:10:34.502 { 00:10:34.502 "name": "Malloc1", 00:10:34.503 "aliases": [ 00:10:34.503 "53f7a2a6-d855-494c-909a-53009eef1eda" 00:10:34.503 ], 00:10:34.503 "product_name": "Malloc disk", 00:10:34.503 "block_size": 512, 00:10:34.503 "num_blocks": 1048576, 00:10:34.503 "uuid": "53f7a2a6-d855-494c-909a-53009eef1eda", 00:10:34.503 "assigned_rate_limits": { 00:10:34.503 "rw_ios_per_sec": 0, 00:10:34.503 "rw_mbytes_per_sec": 0, 00:10:34.503 "r_mbytes_per_sec": 0, 00:10:34.503 "w_mbytes_per_sec": 0 00:10:34.503 }, 00:10:34.503 "claimed": true, 00:10:34.503 "claim_type": "exclusive_write", 00:10:34.503 "zoned": false, 00:10:34.503 "supported_io_types": { 00:10:34.503 "read": true, 00:10:34.503 "write": true, 00:10:34.503 "unmap": true, 00:10:34.503 "flush": true, 00:10:34.503 "reset": true, 00:10:34.503 "nvme_admin": false, 00:10:34.503 "nvme_io": false, 00:10:34.503 "nvme_io_md": false, 00:10:34.503 "write_zeroes": true, 00:10:34.503 "zcopy": true, 00:10:34.503 "get_zone_info": false, 00:10:34.503 "zone_management": false, 00:10:34.503 "zone_append": false, 00:10:34.503 "compare": false, 00:10:34.503 "compare_and_write": false, 00:10:34.503 "abort": true, 00:10:34.503 "seek_hole": false, 00:10:34.503 "seek_data": false, 00:10:34.503 "copy": true, 00:10:34.503 "nvme_iov_md": false 00:10:34.503 }, 00:10:34.503 "memory_domains": [ 00:10:34.503 { 00:10:34.503 "dma_device_id": "system", 00:10:34.503 "dma_device_type": 1 00:10:34.503 }, 00:10:34.503 { 00:10:34.503 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.503 "dma_device_type": 2 00:10:34.503 } 00:10:34.503 ], 00:10:34.503 "driver_specific": {} 00:10:34.503 } 00:10:34.503 ]' 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # jq '.[] .block_size' 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # bs=512 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # jq '.[] .num_blocks' 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # nb=1048576 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1387 -- # bdev_size=512 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1388 -- # echo 512 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:10:34.503 17:31:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:10:36.423 17:31:43 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:10:36.423 17:31:43 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1198 -- # local i=0 00:10:36.423 17:31:43 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:10:36.423 17:31:43 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:10:36.423 17:31:43 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1205 -- # sleep 2 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1208 -- # return 0 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:10:38.343 17:31:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:10:38.343 17:31:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:10:38.343 17:31:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:10:39.283 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@76 -- # '[' 4096 -eq 0 ']' 00:10:39.283 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@81 -- # run_test filesystem_in_capsule_ext4 nvmf_filesystem_create ext4 nvme0n1 00:10:39.283 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:39.283 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:39.283 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:39.553 ************************************ 00:10:39.553 START TEST filesystem_in_capsule_ext4 00:10:39.553 ************************************ 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create ext4 nvme0n1 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@926 -- # local fstype=ext4 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@928 -- # local i=0 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@929 -- # local force 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@931 -- # '[' ext4 = ext4 ']' 00:10:39.553 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@932 -- # force=-F 00:10:39.554 17:31:47 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@937 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:10:39.554 mke2fs 1.47.0 (5-Feb-2023) 00:10:39.554 Discarding device blocks: 0/522240 done 00:10:39.554 Creating filesystem with 522240 1k blocks and 130560 inodes 00:10:39.554 Filesystem UUID: 83cf3cb2-7007-44b6-a070-f7ae7f844f64 00:10:39.554 Superblock backups stored on blocks: 00:10:39.554 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:10:39.554 00:10:39.554 Allocating group tables: 0/64 done 00:10:39.554 Writing inode tables: 0/64 done 00:10:42.857 Creating journal (8192 blocks): done 00:10:44.627 Writing superblocks and filesystem accounting information: 0/6410/64 done 00:10:44.627 00:10:44.627 17:31:52 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@945 -- # return 0 00:10:44.627 17:31:52 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:49.913 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@25 -- # sync 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@27 -- # sync 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@29 -- # i=0 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@37 -- # kill -0 4123804 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:50.175 00:10:50.175 real 0m10.685s 00:10:50.175 user 0m0.030s 00:10:50.175 sys 0m0.081s 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@10 -- # set +x 00:10:50.175 ************************************ 00:10:50.175 END TEST filesystem_in_capsule_ext4 00:10:50.175 ************************************ 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@82 -- # run_test filesystem_in_capsule_btrfs nvmf_filesystem_create btrfs nvme0n1 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:50.175 17:31:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:50.175 ************************************ 00:10:50.175 START TEST filesystem_in_capsule_btrfs 00:10:50.175 ************************************ 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create btrfs nvme0n1 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@926 -- # local fstype=btrfs 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@928 -- # local i=0 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@929 -- # local force 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@931 -- # '[' btrfs = ext4 ']' 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@934 -- # force=-f 00:10:50.175 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@937 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:10:50.436 btrfs-progs v6.8.1 00:10:50.436 See https://btrfs.readthedocs.io for more information. 00:10:50.436 00:10:50.436 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:10:50.436 NOTE: several default settings have changed in version 5.15, please make sure 00:10:50.436 this does not affect your deployments: 00:10:50.436 - DUP for metadata (-m dup) 00:10:50.436 - enabled no-holes (-O no-holes) 00:10:50.436 - enabled free-space-tree (-R free-space-tree) 00:10:50.436 00:10:50.436 Label: (null) 00:10:50.436 UUID: 61875faf-94fc-4ccf-ad65-44cd24c51129 00:10:50.436 Node size: 16384 00:10:50.436 Sector size: 4096 (CPU page size: 4096) 00:10:50.436 Filesystem size: 510.00MiB 00:10:50.436 Block group profiles: 00:10:50.436 Data: single 8.00MiB 00:10:50.436 Metadata: DUP 32.00MiB 00:10:50.436 System: DUP 8.00MiB 00:10:50.436 SSD detected: yes 00:10:50.436 Zoned device: no 00:10:50.436 Features: extref, skinny-metadata, no-holes, free-space-tree 00:10:50.436 Checksum: crc32c 00:10:50.436 Number of devices: 1 00:10:50.436 Devices: 00:10:50.436 ID SIZE PATH 00:10:50.436 1 510.00MiB /dev/nvme0n1p1 00:10:50.436 00:10:50.436 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@945 -- # return 0 00:10:50.436 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@25 -- # sync 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@27 -- # sync 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@29 -- # i=0 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@37 -- # kill -0 4123804 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:50.710 00:10:50.710 real 0m0.567s 00:10:50.710 user 0m0.029s 00:10:50.710 sys 0m0.116s 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@10 -- # set +x 00:10:50.710 ************************************ 00:10:50.710 END TEST filesystem_in_capsule_btrfs 00:10:50.710 ************************************ 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@83 -- # run_test filesystem_in_capsule_xfs nvmf_filesystem_create xfs nvme0n1 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:50.710 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:50.972 ************************************ 00:10:50.972 START TEST filesystem_in_capsule_xfs 00:10:50.972 ************************************ 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create xfs nvme0n1 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@926 -- # local fstype=xfs 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@928 -- # local i=0 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@929 -- # local force 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@931 -- # '[' xfs = ext4 ']' 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@934 -- # force=-f 00:10:50.972 17:31:58 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@937 -- # mkfs.xfs -f /dev/nvme0n1p1 00:10:50.972 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:10:50.972 = sectsz=512 attr=2, projid32bit=1 00:10:50.972 = crc=1 finobt=1, sparse=1, rmapbt=0 00:10:50.972 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:10:50.972 data = bsize=4096 blocks=130560, imaxpct=25 00:10:50.972 = sunit=0 swidth=0 blks 00:10:50.972 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:10:50.972 log =internal log bsize=4096 blocks=16384, version=2 00:10:50.972 = sectsz=512 sunit=0 blks, lazy-count=1 00:10:50.972 realtime =none extsz=4096 blocks=0, rtextents=0 00:10:51.914 Discarding blocks...Done. 00:10:51.914 17:31:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@945 -- # return 0 00:10:51.914 17:31:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@25 -- # sync 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@27 -- # sync 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@29 -- # i=0 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@37 -- # kill -0 4123804 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:53.824 00:10:53.824 real 0m2.859s 00:10:53.824 user 0m0.021s 00:10:53.824 sys 0m0.081s 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@10 -- # set +x 00:10:53.824 ************************************ 00:10:53.824 END TEST filesystem_in_capsule_xfs 00:10:53.824 ************************************ 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@93 -- # sync 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:53.824 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1219 -- # local i=0 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:10:53.824 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1231 -- # return 0 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@101 -- # killprocess 4123804 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@950 -- # '[' -z 4123804 ']' 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@954 -- # kill -0 4123804 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@955 -- # uname 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4123804 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4123804' 00:10:54.085 killing process with pid 4123804 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@969 -- # kill 4123804 00:10:54.085 17:32:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@974 -- # wait 4123804 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:10:54.345 00:10:54.345 real 0m20.861s 00:10:54.345 user 1m22.613s 00:10:54.345 sys 0m1.408s 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:10:54.345 ************************************ 00:10:54.345 END TEST nvmf_filesystem_in_capsule 00:10:54.345 ************************************ 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@108 -- # nvmftestfini 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@514 -- # nvmfcleanup 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@121 -- # sync 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@124 -- # set +e 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:54.345 rmmod nvme_tcp 00:10:54.345 rmmod nvme_fabrics 00:10:54.345 rmmod nvme_keyring 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@128 -- # set -e 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@129 -- # return 0 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@297 -- # iptr 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@789 -- # iptables-save 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@789 -- # iptables-restore 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:54.345 17:32:02 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:56.888 00:10:56.888 real 0m51.613s 00:10:56.888 user 2m45.153s 00:10:56.888 sys 0m8.847s 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:10:56.888 ************************************ 00:10:56.888 END TEST nvmf_filesystem 00:10:56.888 ************************************ 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@18 -- # run_test nvmf_target_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:10:56.888 ************************************ 00:10:56.888 START TEST nvmf_target_discovery 00:10:56.888 ************************************ 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:10:56.888 * Looking for test storage... 00:10:56.888 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1691 -- # lcov --version 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@344 -- # case "$op" in 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@345 -- # : 1 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:56.888 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # decimal 1 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=1 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 1 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # decimal 2 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=2 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 2 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # return 0 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:10:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.889 --rc genhtml_branch_coverage=1 00:10:56.889 --rc genhtml_function_coverage=1 00:10:56.889 --rc genhtml_legend=1 00:10:56.889 --rc geninfo_all_blocks=1 00:10:56.889 --rc geninfo_unexecuted_blocks=1 00:10:56.889 00:10:56.889 ' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:10:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.889 --rc genhtml_branch_coverage=1 00:10:56.889 --rc genhtml_function_coverage=1 00:10:56.889 --rc genhtml_legend=1 00:10:56.889 --rc geninfo_all_blocks=1 00:10:56.889 --rc geninfo_unexecuted_blocks=1 00:10:56.889 00:10:56.889 ' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:10:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.889 --rc genhtml_branch_coverage=1 00:10:56.889 --rc genhtml_function_coverage=1 00:10:56.889 --rc genhtml_legend=1 00:10:56.889 --rc geninfo_all_blocks=1 00:10:56.889 --rc geninfo_unexecuted_blocks=1 00:10:56.889 00:10:56.889 ' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:10:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.889 --rc genhtml_branch_coverage=1 00:10:56.889 --rc genhtml_function_coverage=1 00:10:56.889 --rc genhtml_legend=1 00:10:56.889 --rc geninfo_all_blocks=1 00:10:56.889 --rc geninfo_unexecuted_blocks=1 00:10:56.889 00:10:56.889 ' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # uname -s 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@5 -- # export PATH 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@51 -- # : 0 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:56.889 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@11 -- # NULL_BDEV_SIZE=102400 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@12 -- # NULL_BLOCK_SIZE=512 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@13 -- # NVMF_PORT_REFERRAL=4430 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@15 -- # hash nvme 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@20 -- # nvmftestinit 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@474 -- # prepare_net_devs 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@436 -- # local -g is_hw=no 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@438 -- # remove_spdk_ns 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:10:56.889 17:32:04 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # e810=() 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # x722=() 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # mlx=() 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:11:05.028 Found 0000:31:00.0 (0x8086 - 0x159b) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:11:05.028 Found 0000:31:00.1 (0x8086 - 0x159b) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:11:05.028 Found net devices under 0000:31:00.0: cvl_0_0 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:05.028 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:11:05.029 Found net devices under 0000:31:00.1: cvl_0_1 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # is_hw=yes 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:05.029 17:32:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:05.029 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:05.029 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.501 ms 00:11:05.029 00:11:05.029 --- 10.0.0.2 ping statistics --- 00:11:05.029 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:05.029 rtt min/avg/max/mdev = 0.501/0.501/0.501/0.000 ms 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:05.029 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:05.029 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.322 ms 00:11:05.029 00:11:05.029 --- 10.0.0.1 ping statistics --- 00:11:05.029 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:05.029 rtt min/avg/max/mdev = 0.322/0.322/0.322/0.000 ms 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@448 -- # return 0 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@21 -- # nvmfappstart -m 0xF 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@724 -- # xtrace_disable 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@507 -- # nvmfpid=4132409 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@508 -- # waitforlisten 4132409 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@831 -- # '[' -z 4132409 ']' 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:05.029 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:05.029 17:32:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.029 [2024-10-17 17:32:12.314895] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:11:05.029 [2024-10-17 17:32:12.314958] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:05.029 [2024-10-17 17:32:12.403676] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:05.029 [2024-10-17 17:32:12.456775] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:05.029 [2024-10-17 17:32:12.456824] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:05.029 [2024-10-17 17:32:12.456833] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:05.029 [2024-10-17 17:32:12.456840] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:05.029 [2024-10-17 17:32:12.456847] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:05.029 [2024-10-17 17:32:12.458982] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:05.029 [2024-10-17 17:32:12.459123] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:05.029 [2024-10-17 17:32:12.459281] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:05.029 [2024-10-17 17:32:12.459282] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@864 -- # return 0 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@730 -- # xtrace_disable 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.290 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.290 [2024-10-17 17:32:13.203462] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # seq 1 4 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null1 102400 512 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 Null1 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Null1 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 [2024-10-17 17:32:13.263940] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null2 102400 512 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 Null2 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Null2 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null3 102400 512 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 Null3 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK00000000000003 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 Null3 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:11:05.551 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null4 102400 512 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.552 Null4 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode4 -a -s SPDK00000000000004 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode4 Null4 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode4 -t tcp -a 10.0.0.2 -s 4420 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@32 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@35 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 10.0.0.2 -s 4430 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.552 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@37 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 4420 00:11:05.814 00:11:05.814 Discovery Log Number of Records 6, Generation counter 6 00:11:05.814 =====Discovery Log Entry 0====== 00:11:05.814 trtype: tcp 00:11:05.814 adrfam: ipv4 00:11:05.814 subtype: current discovery subsystem 00:11:05.814 treq: not required 00:11:05.814 portid: 0 00:11:05.814 trsvcid: 4420 00:11:05.814 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:11:05.814 traddr: 10.0.0.2 00:11:05.814 eflags: explicit discovery connections, duplicate discovery information 00:11:05.814 sectype: none 00:11:05.814 =====Discovery Log Entry 1====== 00:11:05.814 trtype: tcp 00:11:05.814 adrfam: ipv4 00:11:05.814 subtype: nvme subsystem 00:11:05.814 treq: not required 00:11:05.814 portid: 0 00:11:05.814 trsvcid: 4420 00:11:05.814 subnqn: nqn.2016-06.io.spdk:cnode1 00:11:05.814 traddr: 10.0.0.2 00:11:05.814 eflags: none 00:11:05.814 sectype: none 00:11:05.814 =====Discovery Log Entry 2====== 00:11:05.814 trtype: tcp 00:11:05.814 adrfam: ipv4 00:11:05.814 subtype: nvme subsystem 00:11:05.814 treq: not required 00:11:05.814 portid: 0 00:11:05.814 trsvcid: 4420 00:11:05.814 subnqn: nqn.2016-06.io.spdk:cnode2 00:11:05.814 traddr: 10.0.0.2 00:11:05.814 eflags: none 00:11:05.814 sectype: none 00:11:05.814 =====Discovery Log Entry 3====== 00:11:05.814 trtype: tcp 00:11:05.814 adrfam: ipv4 00:11:05.814 subtype: nvme subsystem 00:11:05.814 treq: not required 00:11:05.814 portid: 0 00:11:05.814 trsvcid: 4420 00:11:05.814 subnqn: nqn.2016-06.io.spdk:cnode3 00:11:05.814 traddr: 10.0.0.2 00:11:05.814 eflags: none 00:11:05.814 sectype: none 00:11:05.814 =====Discovery Log Entry 4====== 00:11:05.814 trtype: tcp 00:11:05.814 adrfam: ipv4 00:11:05.814 subtype: nvme subsystem 00:11:05.814 treq: not required 00:11:05.814 portid: 0 00:11:05.814 trsvcid: 4420 00:11:05.814 subnqn: nqn.2016-06.io.spdk:cnode4 00:11:05.814 traddr: 10.0.0.2 00:11:05.814 eflags: none 00:11:05.814 sectype: none 00:11:05.814 =====Discovery Log Entry 5====== 00:11:05.814 trtype: tcp 00:11:05.814 adrfam: ipv4 00:11:05.814 subtype: discovery subsystem referral 00:11:05.814 treq: not required 00:11:05.814 portid: 0 00:11:05.814 trsvcid: 4430 00:11:05.814 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:11:05.814 traddr: 10.0.0.2 00:11:05.814 eflags: none 00:11:05.814 sectype: none 00:11:05.814 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@39 -- # echo 'Perform nvmf subsystem discovery via RPC' 00:11:05.814 Perform nvmf subsystem discovery via RPC 00:11:05.814 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@40 -- # rpc_cmd nvmf_get_subsystems 00:11:05.814 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.814 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.814 [ 00:11:05.814 { 00:11:05.814 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:11:05.814 "subtype": "Discovery", 00:11:05.814 "listen_addresses": [ 00:11:05.814 { 00:11:05.814 "trtype": "TCP", 00:11:05.814 "adrfam": "IPv4", 00:11:05.814 "traddr": "10.0.0.2", 00:11:05.814 "trsvcid": "4420" 00:11:05.814 } 00:11:05.814 ], 00:11:05.814 "allow_any_host": true, 00:11:05.814 "hosts": [] 00:11:05.814 }, 00:11:05.814 { 00:11:05.814 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:11:05.814 "subtype": "NVMe", 00:11:05.814 "listen_addresses": [ 00:11:05.814 { 00:11:05.814 "trtype": "TCP", 00:11:05.814 "adrfam": "IPv4", 00:11:05.814 "traddr": "10.0.0.2", 00:11:05.814 "trsvcid": "4420" 00:11:05.814 } 00:11:05.814 ], 00:11:05.814 "allow_any_host": true, 00:11:05.814 "hosts": [], 00:11:05.814 "serial_number": "SPDK00000000000001", 00:11:05.814 "model_number": "SPDK bdev Controller", 00:11:05.815 "max_namespaces": 32, 00:11:05.815 "min_cntlid": 1, 00:11:05.815 "max_cntlid": 65519, 00:11:05.815 "namespaces": [ 00:11:05.815 { 00:11:05.815 "nsid": 1, 00:11:05.815 "bdev_name": "Null1", 00:11:05.815 "name": "Null1", 00:11:05.815 "nguid": "87DF58068C16480EA85DFD0C43FE98D8", 00:11:05.815 "uuid": "87df5806-8c16-480e-a85d-fd0c43fe98d8" 00:11:05.815 } 00:11:05.815 ] 00:11:05.815 }, 00:11:05.815 { 00:11:05.815 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:11:05.815 "subtype": "NVMe", 00:11:05.815 "listen_addresses": [ 00:11:05.815 { 00:11:05.815 "trtype": "TCP", 00:11:05.815 "adrfam": "IPv4", 00:11:05.815 "traddr": "10.0.0.2", 00:11:05.815 "trsvcid": "4420" 00:11:05.815 } 00:11:05.815 ], 00:11:05.815 "allow_any_host": true, 00:11:05.815 "hosts": [], 00:11:05.815 "serial_number": "SPDK00000000000002", 00:11:05.815 "model_number": "SPDK bdev Controller", 00:11:05.815 "max_namespaces": 32, 00:11:05.815 "min_cntlid": 1, 00:11:05.815 "max_cntlid": 65519, 00:11:05.815 "namespaces": [ 00:11:05.815 { 00:11:05.815 "nsid": 1, 00:11:05.815 "bdev_name": "Null2", 00:11:05.815 "name": "Null2", 00:11:05.815 "nguid": "96215D3FD7D746198E611631898A9F80", 00:11:05.815 "uuid": "96215d3f-d7d7-4619-8e61-1631898a9f80" 00:11:05.815 } 00:11:05.815 ] 00:11:05.815 }, 00:11:05.815 { 00:11:05.815 "nqn": "nqn.2016-06.io.spdk:cnode3", 00:11:05.815 "subtype": "NVMe", 00:11:05.815 "listen_addresses": [ 00:11:05.815 { 00:11:05.815 "trtype": "TCP", 00:11:05.815 "adrfam": "IPv4", 00:11:05.815 "traddr": "10.0.0.2", 00:11:05.815 "trsvcid": "4420" 00:11:05.815 } 00:11:05.815 ], 00:11:05.815 "allow_any_host": true, 00:11:05.815 "hosts": [], 00:11:05.815 "serial_number": "SPDK00000000000003", 00:11:05.815 "model_number": "SPDK bdev Controller", 00:11:05.815 "max_namespaces": 32, 00:11:05.815 "min_cntlid": 1, 00:11:05.815 "max_cntlid": 65519, 00:11:05.815 "namespaces": [ 00:11:05.815 { 00:11:05.815 "nsid": 1, 00:11:05.815 "bdev_name": "Null3", 00:11:05.815 "name": "Null3", 00:11:05.815 "nguid": "D7EA470D02A84B1FBB0D612BFCD195F0", 00:11:05.815 "uuid": "d7ea470d-02a8-4b1f-bb0d-612bfcd195f0" 00:11:05.815 } 00:11:05.815 ] 00:11:05.815 }, 00:11:05.815 { 00:11:05.815 "nqn": "nqn.2016-06.io.spdk:cnode4", 00:11:05.815 "subtype": "NVMe", 00:11:05.815 "listen_addresses": [ 00:11:05.815 { 00:11:05.815 "trtype": "TCP", 00:11:05.815 "adrfam": "IPv4", 00:11:05.815 "traddr": "10.0.0.2", 00:11:05.815 "trsvcid": "4420" 00:11:05.815 } 00:11:05.815 ], 00:11:05.815 "allow_any_host": true, 00:11:05.815 "hosts": [], 00:11:05.815 "serial_number": "SPDK00000000000004", 00:11:05.815 "model_number": "SPDK bdev Controller", 00:11:05.815 "max_namespaces": 32, 00:11:05.815 "min_cntlid": 1, 00:11:05.815 "max_cntlid": 65519, 00:11:05.815 "namespaces": [ 00:11:05.815 { 00:11:05.815 "nsid": 1, 00:11:05.815 "bdev_name": "Null4", 00:11:05.815 "name": "Null4", 00:11:05.815 "nguid": "38590ABD41E74DEF8DD7165146EC2000", 00:11:05.815 "uuid": "38590abd-41e7-4def-8dd7-165146ec2000" 00:11:05.815 } 00:11:05.815 ] 00:11:05.815 } 00:11:05.815 ] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # seq 1 4 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null1 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null2 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null3 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode4 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null4 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@47 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 10.0.0.2 -s 4430 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # rpc_cmd bdev_get_bdevs 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # jq -r '.[].name' 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.815 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # check_bdevs= 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@50 -- # '[' -n '' ']' 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@55 -- # trap - SIGINT SIGTERM EXIT 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@57 -- # nvmftestfini 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@514 -- # nvmfcleanup 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@121 -- # sync 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@124 -- # set +e 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:06.075 rmmod nvme_tcp 00:11:06.075 rmmod nvme_fabrics 00:11:06.075 rmmod nvme_keyring 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@128 -- # set -e 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@129 -- # return 0 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@515 -- # '[' -n 4132409 ']' 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@516 -- # killprocess 4132409 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@950 -- # '[' -z 4132409 ']' 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@954 -- # kill -0 4132409 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@955 -- # uname 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:06.075 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4132409 00:11:06.076 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:06.076 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:06.076 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4132409' 00:11:06.076 killing process with pid 4132409 00:11:06.076 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@969 -- # kill 4132409 00:11:06.076 17:32:13 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@974 -- # wait 4132409 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@297 -- # iptr 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@789 -- # iptables-save 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@789 -- # iptables-restore 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:06.336 17:32:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:08.248 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:08.248 00:11:08.248 real 0m11.826s 00:11:08.248 user 0m8.865s 00:11:08.248 sys 0m6.191s 00:11:08.248 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:08.248 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:11:08.248 ************************************ 00:11:08.248 END TEST nvmf_target_discovery 00:11:08.248 ************************************ 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@19 -- # run_test nvmf_referrals /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:11:08.509 ************************************ 00:11:08.509 START TEST nvmf_referrals 00:11:08.509 ************************************ 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:11:08.509 * Looking for test storage... 00:11:08.509 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1691 -- # lcov --version 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # IFS=.-: 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # read -ra ver1 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # IFS=.-: 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # read -ra ver2 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@338 -- # local 'op=<' 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@340 -- # ver1_l=2 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@341 -- # ver2_l=1 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@344 -- # case "$op" in 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@345 -- # : 1 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:08.509 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # decimal 1 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=1 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 1 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # ver1[v]=1 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # decimal 2 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=2 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 2 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # ver2[v]=2 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # return 0 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:11:08.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:08.771 --rc genhtml_branch_coverage=1 00:11:08.771 --rc genhtml_function_coverage=1 00:11:08.771 --rc genhtml_legend=1 00:11:08.771 --rc geninfo_all_blocks=1 00:11:08.771 --rc geninfo_unexecuted_blocks=1 00:11:08.771 00:11:08.771 ' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:11:08.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:08.771 --rc genhtml_branch_coverage=1 00:11:08.771 --rc genhtml_function_coverage=1 00:11:08.771 --rc genhtml_legend=1 00:11:08.771 --rc geninfo_all_blocks=1 00:11:08.771 --rc geninfo_unexecuted_blocks=1 00:11:08.771 00:11:08.771 ' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:11:08.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:08.771 --rc genhtml_branch_coverage=1 00:11:08.771 --rc genhtml_function_coverage=1 00:11:08.771 --rc genhtml_legend=1 00:11:08.771 --rc geninfo_all_blocks=1 00:11:08.771 --rc geninfo_unexecuted_blocks=1 00:11:08.771 00:11:08.771 ' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:11:08.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:08.771 --rc genhtml_branch_coverage=1 00:11:08.771 --rc genhtml_function_coverage=1 00:11:08.771 --rc genhtml_legend=1 00:11:08.771 --rc geninfo_all_blocks=1 00:11:08.771 --rc geninfo_unexecuted_blocks=1 00:11:08.771 00:11:08.771 ' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # uname -s 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@15 -- # shopt -s extglob 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@5 -- # export PATH 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@51 -- # : 0 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:08.771 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@11 -- # NVMF_REFERRAL_IP_1=127.0.0.2 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@12 -- # NVMF_REFERRAL_IP_2=127.0.0.3 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@13 -- # NVMF_REFERRAL_IP_3=127.0.0.4 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@14 -- # NVMF_PORT_REFERRAL=4430 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@15 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@16 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@37 -- # nvmftestinit 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:08.771 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@474 -- # prepare_net_devs 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@436 -- # local -g is_hw=no 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@438 -- # remove_spdk_ns 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@309 -- # xtrace_disable 00:11:08.772 17:32:16 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # pci_devs=() 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # net_devs=() 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # e810=() 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # local -ga e810 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # x722=() 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # local -ga x722 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # mlx=() 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # local -ga mlx 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:17.015 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:11:17.016 Found 0000:31:00.0 (0x8086 - 0x159b) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:11:17.016 Found 0000:31:00.1 (0x8086 - 0x159b) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:11:17.016 Found net devices under 0000:31:00.0: cvl_0_0 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:11:17.016 Found net devices under 0000:31:00.1: cvl_0_1 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # is_hw=yes 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:17.016 17:32:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:17.016 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:17.016 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.683 ms 00:11:17.016 00:11:17.016 --- 10.0.0.2 ping statistics --- 00:11:17.016 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:17.016 rtt min/avg/max/mdev = 0.683/0.683/0.683/0.000 ms 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:17.016 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:17.016 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.276 ms 00:11:17.016 00:11:17.016 --- 10.0.0.1 ping statistics --- 00:11:17.016 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:17.016 rtt min/avg/max/mdev = 0.276/0.276/0.276/0.000 ms 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@448 -- # return 0 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@38 -- # nvmfappstart -m 0xF 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@724 -- # xtrace_disable 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@507 -- # nvmfpid=4136924 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@508 -- # waitforlisten 4136924 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@831 -- # '[' -z 4136924 ']' 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:17.016 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:17.016 17:32:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.016 [2024-10-17 17:32:24.177007] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:11:17.016 [2024-10-17 17:32:24.177077] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:17.016 [2024-10-17 17:32:24.266914] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:17.016 [2024-10-17 17:32:24.321659] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:17.016 [2024-10-17 17:32:24.321723] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:17.016 [2024-10-17 17:32:24.321732] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:17.016 [2024-10-17 17:32:24.321739] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:17.016 [2024-10-17 17:32:24.321746] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:17.016 [2024-10-17 17:32:24.324151] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:17.016 [2024-10-17 17:32:24.324316] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:17.016 [2024-10-17 17:32:24.324475] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:17.016 [2024-10-17 17:32:24.324475] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:17.277 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:17.277 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@864 -- # return 0 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@730 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@40 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 [2024-10-17 17:32:25.060995] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 10.0.0.2 -s 8009 discovery 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 [2024-10-17 17:32:25.077404] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@44 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@45 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.3 -s 4430 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@46 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.4 -s 4430 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # rpc_cmd nvmf_discovery_get_referrals 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # jq length 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # (( 3 == 3 )) 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # get_referral_ips rpc 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:11:17.278 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # get_referral_ips nvme 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:11:17.538 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@52 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@53 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.3 -s 4430 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@54 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.4 -s 4430 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # rpc_cmd nvmf_discovery_get_referrals 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # jq length 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # (( 0 == 0 )) 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # get_referral_ips nvme 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:11:17.799 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:11:18.060 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:11:18.060 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # [[ '' == '' ]] 00:11:18.060 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@60 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n discovery 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@62 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # get_referral_ips rpc 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.2 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # get_referral_ips nvme 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:11:18.061 17:32:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:11:18.322 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.2 00:11:18.322 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:11:18.322 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # get_discovery_entries 'nvme subsystem' 00:11:18.322 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # jq -r .subnqn 00:11:18.322 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:11:18.322 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:18.323 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:11:18.584 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # [[ nqn.2016-06.io.spdk:cnode1 == \n\q\n\.\2\0\1\6\-\0\6\.\i\o\.\s\p\d\k\:\c\n\o\d\e\1 ]] 00:11:18.584 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # get_discovery_entries 'discovery subsystem referral' 00:11:18.584 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # jq -r .subnqn 00:11:18.584 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:11:18.585 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:18.585 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@71 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # get_referral_ips rpc 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # get_referral_ips nvme 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:11:18.846 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # get_discovery_entries 'nvme subsystem' 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # jq -r .subnqn 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # [[ '' == '' ]] 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # get_discovery_entries 'discovery subsystem referral' 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # jq -r .subnqn 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:19.106 17:32:26 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@79 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2014-08.org.nvmexpress.discovery 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # rpc_cmd nvmf_discovery_get_referrals 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # jq length 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # (( 0 == 0 )) 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # get_referral_ips nvme 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 8009 -o json 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:11:19.367 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # [[ '' == '' ]] 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@85 -- # trap - SIGINT SIGTERM EXIT 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@86 -- # nvmftestfini 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@514 -- # nvmfcleanup 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@121 -- # sync 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@124 -- # set +e 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:19.627 rmmod nvme_tcp 00:11:19.627 rmmod nvme_fabrics 00:11:19.627 rmmod nvme_keyring 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@128 -- # set -e 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@129 -- # return 0 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@515 -- # '[' -n 4136924 ']' 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@516 -- # killprocess 4136924 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@950 -- # '[' -z 4136924 ']' 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@954 -- # kill -0 4136924 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@955 -- # uname 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4136924 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4136924' 00:11:19.627 killing process with pid 4136924 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@969 -- # kill 4136924 00:11:19.627 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@974 -- # wait 4136924 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@297 -- # iptr 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@789 -- # iptables-save 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@789 -- # iptables-restore 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:19.888 17:32:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:21.801 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:21.801 00:11:21.801 real 0m13.461s 00:11:21.801 user 0m16.302s 00:11:21.801 sys 0m6.650s 00:11:21.801 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:21.801 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:11:21.801 ************************************ 00:11:21.801 END TEST nvmf_referrals 00:11:21.801 ************************************ 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@20 -- # run_test nvmf_connect_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:11:22.062 ************************************ 00:11:22.062 START TEST nvmf_connect_disconnect 00:11:22.062 ************************************ 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:11:22.062 * Looking for test storage... 00:11:22.062 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1691 -- # lcov --version 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@345 -- # : 1 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # decimal 1 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=1 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 1 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:11:22.062 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # decimal 2 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=2 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 2 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # return 0 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:11:22.324 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:22.324 --rc genhtml_branch_coverage=1 00:11:22.324 --rc genhtml_function_coverage=1 00:11:22.324 --rc genhtml_legend=1 00:11:22.324 --rc geninfo_all_blocks=1 00:11:22.324 --rc geninfo_unexecuted_blocks=1 00:11:22.324 00:11:22.324 ' 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:11:22.324 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:22.324 --rc genhtml_branch_coverage=1 00:11:22.324 --rc genhtml_function_coverage=1 00:11:22.324 --rc genhtml_legend=1 00:11:22.324 --rc geninfo_all_blocks=1 00:11:22.324 --rc geninfo_unexecuted_blocks=1 00:11:22.324 00:11:22.324 ' 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:11:22.324 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:22.324 --rc genhtml_branch_coverage=1 00:11:22.324 --rc genhtml_function_coverage=1 00:11:22.324 --rc genhtml_legend=1 00:11:22.324 --rc geninfo_all_blocks=1 00:11:22.324 --rc geninfo_unexecuted_blocks=1 00:11:22.324 00:11:22.324 ' 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:11:22.324 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:22.324 --rc genhtml_branch_coverage=1 00:11:22.324 --rc genhtml_function_coverage=1 00:11:22.324 --rc genhtml_legend=1 00:11:22.324 --rc geninfo_all_blocks=1 00:11:22.324 --rc geninfo_unexecuted_blocks=1 00:11:22.324 00:11:22.324 ' 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # uname -s 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:22.324 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:22.325 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:22.325 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:22.325 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:22.325 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:22.325 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:22.325 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:22.325 17:32:29 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@5 -- # export PATH 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@51 -- # : 0 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:22.325 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@11 -- # MALLOC_BDEV_SIZE=64 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@15 -- # nvmftestinit 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@474 -- # prepare_net_devs 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@436 -- # local -g is_hw=no 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@438 -- # remove_spdk_ns 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:11:22.325 17:32:30 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # e810=() 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # x722=() 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:11:30.463 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:11:30.464 Found 0000:31:00.0 (0x8086 - 0x159b) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:11:30.464 Found 0000:31:00.1 (0x8086 - 0x159b) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:11:30.464 Found net devices under 0000:31:00.0: cvl_0_0 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:11:30.464 Found net devices under 0000:31:00.1: cvl_0_1 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # is_hw=yes 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:30.464 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:30.464 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.634 ms 00:11:30.464 00:11:30.464 --- 10.0.0.2 ping statistics --- 00:11:30.464 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:30.464 rtt min/avg/max/mdev = 0.634/0.634/0.634/0.000 ms 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:30.464 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:30.464 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.326 ms 00:11:30.464 00:11:30.464 --- 10.0.0.1 ping statistics --- 00:11:30.464 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:30.464 rtt min/avg/max/mdev = 0.326/0.326/0.326/0.000 ms 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@448 -- # return 0 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@16 -- # nvmfappstart -m 0xF 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@724 -- # xtrace_disable 00:11:30.464 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@507 -- # nvmfpid=4141957 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@508 -- # waitforlisten 4141957 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@831 -- # '[' -z 4141957 ']' 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:30.465 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:30.465 17:32:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.465 [2024-10-17 17:32:37.769172] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:11:30.465 [2024-10-17 17:32:37.769241] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:30.465 [2024-10-17 17:32:37.858940] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:30.465 [2024-10-17 17:32:37.911697] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:30.465 [2024-10-17 17:32:37.911756] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:30.465 [2024-10-17 17:32:37.911764] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:30.465 [2024-10-17 17:32:37.911771] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:30.465 [2024-10-17 17:32:37.911778] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:30.465 [2024-10-17 17:32:37.913904] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:30.465 [2024-10-17 17:32:37.914066] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:30.465 [2024-10-17 17:32:37.914225] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:30.465 [2024-10-17 17:32:37.914226] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@864 -- # return 0 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@730 -- # xtrace_disable 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.726 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.726 [2024-10-17 17:32:38.638281] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # bdev=Malloc0 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:30.989 [2024-10-17 17:32:38.722708] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@26 -- # '[' 0 -eq 1 ']' 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@31 -- # num_iterations=5 00:11:30.989 17:32:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@34 -- # set +x 00:11:35.193 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:38.491 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:41.791 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:45.997 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:49.299 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@43 -- # trap - SIGINT SIGTERM EXIT 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@45 -- # nvmftestfini 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@514 -- # nvmfcleanup 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@121 -- # sync 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@124 -- # set +e 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:49.299 rmmod nvme_tcp 00:11:49.299 rmmod nvme_fabrics 00:11:49.299 rmmod nvme_keyring 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@128 -- # set -e 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@129 -- # return 0 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@515 -- # '[' -n 4141957 ']' 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@516 -- # killprocess 4141957 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@950 -- # '[' -z 4141957 ']' 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@954 -- # kill -0 4141957 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@955 -- # uname 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4141957 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4141957' 00:11:49.299 killing process with pid 4141957 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@969 -- # kill 4141957 00:11:49.299 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@974 -- # wait 4141957 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@297 -- # iptr 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@789 -- # iptables-save 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@789 -- # iptables-restore 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:49.560 17:32:57 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:51.474 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:51.474 00:11:51.474 real 0m29.580s 00:11:51.474 user 1m19.287s 00:11:51.474 sys 0m7.309s 00:11:51.474 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:51.474 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:11:51.474 ************************************ 00:11:51.474 END TEST nvmf_connect_disconnect 00:11:51.474 ************************************ 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@21 -- # run_test nvmf_multitarget /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:11:51.735 ************************************ 00:11:51.735 START TEST nvmf_multitarget 00:11:51.735 ************************************ 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:11:51.735 * Looking for test storage... 00:11:51.735 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1691 -- # lcov --version 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # IFS=.-: 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # read -ra ver1 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # IFS=.-: 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # read -ra ver2 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@338 -- # local 'op=<' 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@340 -- # ver1_l=2 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@341 -- # ver2_l=1 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@344 -- # case "$op" in 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@345 -- # : 1 00:11:51.735 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # decimal 1 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=1 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 1 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # ver1[v]=1 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # decimal 2 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=2 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:51.736 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 2 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # ver2[v]=2 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # return 0 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:11:51.997 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:51.997 --rc genhtml_branch_coverage=1 00:11:51.997 --rc genhtml_function_coverage=1 00:11:51.997 --rc genhtml_legend=1 00:11:51.997 --rc geninfo_all_blocks=1 00:11:51.997 --rc geninfo_unexecuted_blocks=1 00:11:51.997 00:11:51.997 ' 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:11:51.997 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:51.997 --rc genhtml_branch_coverage=1 00:11:51.997 --rc genhtml_function_coverage=1 00:11:51.997 --rc genhtml_legend=1 00:11:51.997 --rc geninfo_all_blocks=1 00:11:51.997 --rc geninfo_unexecuted_blocks=1 00:11:51.997 00:11:51.997 ' 00:11:51.997 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:11:51.998 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:51.998 --rc genhtml_branch_coverage=1 00:11:51.998 --rc genhtml_function_coverage=1 00:11:51.998 --rc genhtml_legend=1 00:11:51.998 --rc geninfo_all_blocks=1 00:11:51.998 --rc geninfo_unexecuted_blocks=1 00:11:51.998 00:11:51.998 ' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:11:51.998 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:51.998 --rc genhtml_branch_coverage=1 00:11:51.998 --rc genhtml_function_coverage=1 00:11:51.998 --rc genhtml_legend=1 00:11:51.998 --rc geninfo_all_blocks=1 00:11:51.998 --rc geninfo_unexecuted_blocks=1 00:11:51.998 00:11:51.998 ' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # uname -s 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@15 -- # shopt -s extglob 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@5 -- # export PATH 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@51 -- # : 0 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:51.998 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@13 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@15 -- # nvmftestinit 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@474 -- # prepare_net_devs 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@436 -- # local -g is_hw=no 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@438 -- # remove_spdk_ns 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@309 -- # xtrace_disable 00:11:51.998 17:32:59 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # pci_devs=() 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # net_devs=() 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # e810=() 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # local -ga e810 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # x722=() 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # local -ga x722 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # mlx=() 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # local -ga mlx 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:00.147 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:12:00.148 Found 0000:31:00.0 (0x8086 - 0x159b) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:12:00.148 Found 0000:31:00.1 (0x8086 - 0x159b) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:12:00.148 Found net devices under 0000:31:00.0: cvl_0_0 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:12:00.148 Found net devices under 0000:31:00.1: cvl_0_1 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # is_hw=yes 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:00.148 17:33:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:00.148 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:00.148 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.657 ms 00:12:00.148 00:12:00.148 --- 10.0.0.2 ping statistics --- 00:12:00.148 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:00.148 rtt min/avg/max/mdev = 0.657/0.657/0.657/0.000 ms 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:00.148 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:00.148 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.286 ms 00:12:00.148 00:12:00.148 --- 10.0.0.1 ping statistics --- 00:12:00.148 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:00.148 rtt min/avg/max/mdev = 0.286/0.286/0.286/0.000 ms 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@448 -- # return 0 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@16 -- # nvmfappstart -m 0xF 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@724 -- # xtrace_disable 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@507 -- # nvmfpid=4150220 00:12:00.148 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@508 -- # waitforlisten 4150220 00:12:00.149 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:12:00.149 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@831 -- # '[' -z 4150220 ']' 00:12:00.149 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:00.149 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:00.149 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:00.149 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:00.149 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:00.149 17:33:07 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:00.149 [2024-10-17 17:33:07.389840] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:12:00.149 [2024-10-17 17:33:07.389903] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:00.149 [2024-10-17 17:33:07.480919] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:00.149 [2024-10-17 17:33:07.533971] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:00.149 [2024-10-17 17:33:07.534018] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:00.149 [2024-10-17 17:33:07.534026] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:00.149 [2024-10-17 17:33:07.534034] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:00.149 [2024-10-17 17:33:07.534040] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:00.149 [2024-10-17 17:33:07.536134] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:00.149 [2024-10-17 17:33:07.536296] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:12:00.149 [2024-10-17 17:33:07.536462] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:12:00.149 [2024-10-17 17:33:07.536463] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@864 -- # return 0 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@730 -- # xtrace_disable 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@18 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:12:00.410 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # jq length 00:12:00.671 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # '[' 1 '!=' 1 ']' 00:12:00.671 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_1 -s 32 00:12:00.671 "nvmf_tgt_1" 00:12:00.671 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_2 -s 32 00:12:00.671 "nvmf_tgt_2" 00:12:00.932 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:12:00.932 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # jq length 00:12:00.932 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # '[' 3 '!=' 3 ']' 00:12:00.932 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_1 00:12:00.932 true 00:12:00.932 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_2 00:12:01.194 true 00:12:01.194 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:12:01.194 17:33:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # jq length 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # '[' 1 '!=' 1 ']' 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@41 -- # nvmftestfini 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@514 -- # nvmfcleanup 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@121 -- # sync 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@124 -- # set +e 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:01.194 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:01.194 rmmod nvme_tcp 00:12:01.194 rmmod nvme_fabrics 00:12:01.194 rmmod nvme_keyring 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@128 -- # set -e 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@129 -- # return 0 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@515 -- # '[' -n 4150220 ']' 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@516 -- # killprocess 4150220 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@950 -- # '[' -z 4150220 ']' 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@954 -- # kill -0 4150220 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@955 -- # uname 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4150220 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4150220' 00:12:01.455 killing process with pid 4150220 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@969 -- # kill 4150220 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@974 -- # wait 4150220 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@297 -- # iptr 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@789 -- # iptables-save 00:12:01.455 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@789 -- # iptables-restore 00:12:01.716 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:01.716 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:01.717 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:01.717 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:01.717 17:33:09 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:03.633 00:12:03.633 real 0m12.008s 00:12:03.633 user 0m10.402s 00:12:03.633 sys 0m6.228s 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:12:03.633 ************************************ 00:12:03.633 END TEST nvmf_multitarget 00:12:03.633 ************************************ 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@22 -- # run_test nvmf_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:12:03.633 ************************************ 00:12:03.633 START TEST nvmf_rpc 00:12:03.633 ************************************ 00:12:03.633 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:12:03.895 * Looking for test storage... 00:12:03.895 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1691 -- # lcov --version 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:12:03.895 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@344 -- # case "$op" in 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@345 -- # : 1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # decimal 1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # decimal 2 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=2 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 2 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # return 0 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:12:03.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:03.896 --rc genhtml_branch_coverage=1 00:12:03.896 --rc genhtml_function_coverage=1 00:12:03.896 --rc genhtml_legend=1 00:12:03.896 --rc geninfo_all_blocks=1 00:12:03.896 --rc geninfo_unexecuted_blocks=1 00:12:03.896 00:12:03.896 ' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:12:03.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:03.896 --rc genhtml_branch_coverage=1 00:12:03.896 --rc genhtml_function_coverage=1 00:12:03.896 --rc genhtml_legend=1 00:12:03.896 --rc geninfo_all_blocks=1 00:12:03.896 --rc geninfo_unexecuted_blocks=1 00:12:03.896 00:12:03.896 ' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:12:03.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:03.896 --rc genhtml_branch_coverage=1 00:12:03.896 --rc genhtml_function_coverage=1 00:12:03.896 --rc genhtml_legend=1 00:12:03.896 --rc geninfo_all_blocks=1 00:12:03.896 --rc geninfo_unexecuted_blocks=1 00:12:03.896 00:12:03.896 ' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:12:03.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:03.896 --rc genhtml_branch_coverage=1 00:12:03.896 --rc genhtml_function_coverage=1 00:12:03.896 --rc genhtml_legend=1 00:12:03.896 --rc geninfo_all_blocks=1 00:12:03.896 --rc geninfo_unexecuted_blocks=1 00:12:03.896 00:12:03.896 ' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # uname -s 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@15 -- # shopt -s extglob 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@5 -- # export PATH 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@51 -- # : 0 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:03.896 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@11 -- # loops=5 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@23 -- # nvmftestinit 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@474 -- # prepare_net_devs 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@436 -- # local -g is_hw=no 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@438 -- # remove_spdk_ns 00:12:03.896 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:03.897 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:03.897 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:03.897 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:12:03.897 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:12:03.897 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@309 -- # xtrace_disable 00:12:03.897 17:33:11 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # pci_devs=() 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # net_devs=() 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # e810=() 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # local -ga e810 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # x722=() 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # local -ga x722 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # mlx=() 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # local -ga mlx 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:12.042 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:12:12.043 Found 0000:31:00.0 (0x8086 - 0x159b) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:12:12.043 Found 0000:31:00.1 (0x8086 - 0x159b) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:12:12.043 Found net devices under 0000:31:00.0: cvl_0_0 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:12:12.043 Found net devices under 0000:31:00.1: cvl_0_1 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # is_hw=yes 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:12.043 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:12.043 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.591 ms 00:12:12.043 00:12:12.043 --- 10.0.0.2 ping statistics --- 00:12:12.043 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:12.043 rtt min/avg/max/mdev = 0.591/0.591/0.591/0.000 ms 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:12.043 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:12.043 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.272 ms 00:12:12.043 00:12:12.043 --- 10.0.0.1 ping statistics --- 00:12:12.043 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:12.043 rtt min/avg/max/mdev = 0.272/0.272/0.272/0.000 ms 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@448 -- # return 0 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@24 -- # nvmfappstart -m 0xF 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@724 -- # xtrace_disable 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@507 -- # nvmfpid=4155406 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@508 -- # waitforlisten 4155406 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@831 -- # '[' -z 4155406 ']' 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:12.043 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:12.043 17:33:19 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.043 [2024-10-17 17:33:19.509284] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:12:12.043 [2024-10-17 17:33:19.509350] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:12.043 [2024-10-17 17:33:19.598323] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:12.043 [2024-10-17 17:33:19.653599] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:12.043 [2024-10-17 17:33:19.653653] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:12.043 [2024-10-17 17:33:19.653661] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:12.043 [2024-10-17 17:33:19.653668] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:12.043 [2024-10-17 17:33:19.653674] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:12.043 [2024-10-17 17:33:19.655740] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:12.043 [2024-10-17 17:33:19.655834] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:12:12.043 [2024-10-17 17:33:19.655993] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:12:12.044 [2024-10-17 17:33:19.655993] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@864 -- # return 0 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@730 -- # xtrace_disable 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # rpc_cmd nvmf_get_stats 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # stats='{ 00:12:12.616 "tick_rate": 2400000000, 00:12:12.616 "poll_groups": [ 00:12:12.616 { 00:12:12.616 "name": "nvmf_tgt_poll_group_000", 00:12:12.616 "admin_qpairs": 0, 00:12:12.616 "io_qpairs": 0, 00:12:12.616 "current_admin_qpairs": 0, 00:12:12.616 "current_io_qpairs": 0, 00:12:12.616 "pending_bdev_io": 0, 00:12:12.616 "completed_nvme_io": 0, 00:12:12.616 "transports": [] 00:12:12.616 }, 00:12:12.616 { 00:12:12.616 "name": "nvmf_tgt_poll_group_001", 00:12:12.616 "admin_qpairs": 0, 00:12:12.616 "io_qpairs": 0, 00:12:12.616 "current_admin_qpairs": 0, 00:12:12.616 "current_io_qpairs": 0, 00:12:12.616 "pending_bdev_io": 0, 00:12:12.616 "completed_nvme_io": 0, 00:12:12.616 "transports": [] 00:12:12.616 }, 00:12:12.616 { 00:12:12.616 "name": "nvmf_tgt_poll_group_002", 00:12:12.616 "admin_qpairs": 0, 00:12:12.616 "io_qpairs": 0, 00:12:12.616 "current_admin_qpairs": 0, 00:12:12.616 "current_io_qpairs": 0, 00:12:12.616 "pending_bdev_io": 0, 00:12:12.616 "completed_nvme_io": 0, 00:12:12.616 "transports": [] 00:12:12.616 }, 00:12:12.616 { 00:12:12.616 "name": "nvmf_tgt_poll_group_003", 00:12:12.616 "admin_qpairs": 0, 00:12:12.616 "io_qpairs": 0, 00:12:12.616 "current_admin_qpairs": 0, 00:12:12.616 "current_io_qpairs": 0, 00:12:12.616 "pending_bdev_io": 0, 00:12:12.616 "completed_nvme_io": 0, 00:12:12.616 "transports": [] 00:12:12.616 } 00:12:12.616 ] 00:12:12.616 }' 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # jcount '.poll_groups[].name' 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@14 -- # local 'filter=.poll_groups[].name' 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # jq '.poll_groups[].name' 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # wc -l 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # (( 4 == 4 )) 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # jq '.poll_groups[0].transports[0]' 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # [[ null == null ]] 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@31 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.616 [2024-10-17 17:33:20.496621] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # rpc_cmd nvmf_get_stats 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.616 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # stats='{ 00:12:12.616 "tick_rate": 2400000000, 00:12:12.616 "poll_groups": [ 00:12:12.617 { 00:12:12.617 "name": "nvmf_tgt_poll_group_000", 00:12:12.617 "admin_qpairs": 0, 00:12:12.617 "io_qpairs": 0, 00:12:12.617 "current_admin_qpairs": 0, 00:12:12.617 "current_io_qpairs": 0, 00:12:12.617 "pending_bdev_io": 0, 00:12:12.617 "completed_nvme_io": 0, 00:12:12.617 "transports": [ 00:12:12.617 { 00:12:12.617 "trtype": "TCP" 00:12:12.617 } 00:12:12.617 ] 00:12:12.617 }, 00:12:12.617 { 00:12:12.617 "name": "nvmf_tgt_poll_group_001", 00:12:12.617 "admin_qpairs": 0, 00:12:12.617 "io_qpairs": 0, 00:12:12.617 "current_admin_qpairs": 0, 00:12:12.617 "current_io_qpairs": 0, 00:12:12.617 "pending_bdev_io": 0, 00:12:12.617 "completed_nvme_io": 0, 00:12:12.617 "transports": [ 00:12:12.617 { 00:12:12.617 "trtype": "TCP" 00:12:12.617 } 00:12:12.617 ] 00:12:12.617 }, 00:12:12.617 { 00:12:12.617 "name": "nvmf_tgt_poll_group_002", 00:12:12.617 "admin_qpairs": 0, 00:12:12.617 "io_qpairs": 0, 00:12:12.617 "current_admin_qpairs": 0, 00:12:12.617 "current_io_qpairs": 0, 00:12:12.617 "pending_bdev_io": 0, 00:12:12.617 "completed_nvme_io": 0, 00:12:12.617 "transports": [ 00:12:12.617 { 00:12:12.617 "trtype": "TCP" 00:12:12.617 } 00:12:12.617 ] 00:12:12.617 }, 00:12:12.617 { 00:12:12.617 "name": "nvmf_tgt_poll_group_003", 00:12:12.617 "admin_qpairs": 0, 00:12:12.617 "io_qpairs": 0, 00:12:12.617 "current_admin_qpairs": 0, 00:12:12.617 "current_io_qpairs": 0, 00:12:12.617 "pending_bdev_io": 0, 00:12:12.617 "completed_nvme_io": 0, 00:12:12.617 "transports": [ 00:12:12.617 { 00:12:12.617 "trtype": "TCP" 00:12:12.617 } 00:12:12.617 ] 00:12:12.617 } 00:12:12.617 ] 00:12:12.617 }' 00:12:12.617 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # jsum '.poll_groups[].admin_qpairs' 00:12:12.617 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:12:12.617 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:12:12.617 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # (( 0 == 0 )) 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # jsum '.poll_groups[].io_qpairs' 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # (( 0 == 0 )) 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@38 -- # '[' rdma == tcp ']' 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@46 -- # MALLOC_BDEV_SIZE=64 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@47 -- # MALLOC_BLOCK_SIZE=512 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@49 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.879 Malloc1 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@52 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@54 -- # rpc_cmd nvmf_subsystem_allow_any_host -d nqn.2016-06.io.spdk:cnode1 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@55 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.879 [2024-10-17 17:33:20.709039] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@58 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -a 10.0.0.2 -s 4420 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@650 -- # local es=0 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -a 10.0.0.2 -s 4420 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@638 -- # local arg=nvme 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # type -t nvme 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -P nvme 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # arg=/usr/sbin/nvme 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # [[ -x /usr/sbin/nvme ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -a 10.0.0.2 -s 4420 00:12:12.879 [2024-10-17 17:33:20.746299] ctrlr.c: 823:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6' 00:12:12.879 Failed to write to /dev/nvme-fabrics: Input/output error 00:12:12.879 could not add new controller: failed to write to nvme-fabrics device 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # es=1 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@61 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.879 17:33:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@62 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:14.792 17:33:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@63 -- # waitforserial SPDKISFASTANDAWESOME 00:12:14.792 17:33:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:12:14.792 17:33:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:12:14.792 17:33:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:12:14.792 17:33:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:12:16.703 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@64 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:12:16.704 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@65 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@68 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@69 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@650 -- # local es=0 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@638 -- # local arg=nvme 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # type -t nvme 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -P nvme 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # arg=/usr/sbin/nvme 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # [[ -x /usr/sbin/nvme ]] 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:16.704 [2024-10-17 17:33:24.482167] ctrlr.c: 823:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6' 00:12:16.704 Failed to write to /dev/nvme-fabrics: Input/output error 00:12:16.704 could not add new controller: failed to write to nvme-fabrics device 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # es=1 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@72 -- # rpc_cmd nvmf_subsystem_allow_any_host -e nqn.2016-06.io.spdk:cnode1 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.704 17:33:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@73 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:18.615 17:33:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@74 -- # waitforserial SPDKISFASTANDAWESOME 00:12:18.615 17:33:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:12:18.615 17:33:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:12:18.615 17:33:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:12:18.615 17:33:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@75 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:12:20.534 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@76 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@78 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # seq 1 5 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:20.534 [2024-10-17 17:33:28.235524] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.534 17:33:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:21.997 17:33:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:12:21.997 17:33:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:12:21.997 17:33:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:12:21.997 17:33:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:12:21.997 17:33:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:12:23.944 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:12:23.944 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:12:23.945 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:12:23.945 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:12:23.945 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:12:23.945 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:12:23.945 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:12:24.206 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:24.206 [2024-10-17 17:33:31.979656] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.206 17:33:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:24.206 17:33:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.206 17:33:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:25.590 17:33:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:12:25.590 17:33:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:12:25.590 17:33:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:12:25.590 17:33:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:12:25.590 17:33:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:12:28.131 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:28.131 [2024-10-17 17:33:35.691982] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.131 17:33:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:29.512 17:33:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:12:29.512 17:33:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:12:29.512 17:33:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:12:29.512 17:33:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:12:29.512 17:33:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:12:31.421 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:12:31.421 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:12:31.421 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:12:31.421 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:12:31.421 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:12:31.421 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:12:31.421 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:12:31.682 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:31.682 [2024-10-17 17:33:39.446204] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.682 17:33:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:33.064 17:33:40 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:12:33.064 17:33:40 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:12:33.064 17:33:40 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:12:33.064 17:33:40 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:12:33.064 17:33:40 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:12:35.610 17:33:42 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:12:35.610 17:33:42 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:12:35.610 17:33:42 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:12:35.610 17:33:42 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:12:35.610 17:33:42 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:12:35.610 17:33:42 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:12:35.610 17:33:42 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:12:35.610 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:35.610 [2024-10-17 17:33:43.154476] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.610 17:33:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:12:36.992 17:33:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:12:36.992 17:33:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:12:36.992 17:33:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:12:36.992 17:33:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:12:36.992 17:33:44 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:12:38.901 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:12:38.901 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:12:38.901 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:12:38.901 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:12:38.901 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:12:38.901 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:12:38.901 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:12:39.163 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # seq 1 5 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 [2024-10-17 17:33:46.908696] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 [2024-10-17 17:33:46.976876] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.163 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.164 [2024-10-17 17:33:47.045073] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.164 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 [2024-10-17 17:33:47.113282] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.425 [2024-10-17 17:33:47.177495] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:12:39.425 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # rpc_cmd nvmf_get_stats 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # stats='{ 00:12:39.426 "tick_rate": 2400000000, 00:12:39.426 "poll_groups": [ 00:12:39.426 { 00:12:39.426 "name": "nvmf_tgt_poll_group_000", 00:12:39.426 "admin_qpairs": 0, 00:12:39.426 "io_qpairs": 224, 00:12:39.426 "current_admin_qpairs": 0, 00:12:39.426 "current_io_qpairs": 0, 00:12:39.426 "pending_bdev_io": 0, 00:12:39.426 "completed_nvme_io": 275, 00:12:39.426 "transports": [ 00:12:39.426 { 00:12:39.426 "trtype": "TCP" 00:12:39.426 } 00:12:39.426 ] 00:12:39.426 }, 00:12:39.426 { 00:12:39.426 "name": "nvmf_tgt_poll_group_001", 00:12:39.426 "admin_qpairs": 1, 00:12:39.426 "io_qpairs": 223, 00:12:39.426 "current_admin_qpairs": 0, 00:12:39.426 "current_io_qpairs": 0, 00:12:39.426 "pending_bdev_io": 0, 00:12:39.426 "completed_nvme_io": 384, 00:12:39.426 "transports": [ 00:12:39.426 { 00:12:39.426 "trtype": "TCP" 00:12:39.426 } 00:12:39.426 ] 00:12:39.426 }, 00:12:39.426 { 00:12:39.426 "name": "nvmf_tgt_poll_group_002", 00:12:39.426 "admin_qpairs": 6, 00:12:39.426 "io_qpairs": 218, 00:12:39.426 "current_admin_qpairs": 0, 00:12:39.426 "current_io_qpairs": 0, 00:12:39.426 "pending_bdev_io": 0, 00:12:39.426 "completed_nvme_io": 351, 00:12:39.426 "transports": [ 00:12:39.426 { 00:12:39.426 "trtype": "TCP" 00:12:39.426 } 00:12:39.426 ] 00:12:39.426 }, 00:12:39.426 { 00:12:39.426 "name": "nvmf_tgt_poll_group_003", 00:12:39.426 "admin_qpairs": 0, 00:12:39.426 "io_qpairs": 224, 00:12:39.426 "current_admin_qpairs": 0, 00:12:39.426 "current_io_qpairs": 0, 00:12:39.426 "pending_bdev_io": 0, 00:12:39.426 "completed_nvme_io": 229, 00:12:39.426 "transports": [ 00:12:39.426 { 00:12:39.426 "trtype": "TCP" 00:12:39.426 } 00:12:39.426 ] 00:12:39.426 } 00:12:39.426 ] 00:12:39.426 }' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # jsum '.poll_groups[].admin_qpairs' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # (( 7 > 0 )) 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # jsum '.poll_groups[].io_qpairs' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # (( 889 > 0 )) 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@115 -- # '[' rdma == tcp ']' 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@123 -- # nvmftestfini 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@514 -- # nvmfcleanup 00:12:39.426 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@121 -- # sync 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@124 -- # set +e 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:39.688 rmmod nvme_tcp 00:12:39.688 rmmod nvme_fabrics 00:12:39.688 rmmod nvme_keyring 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@128 -- # set -e 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@129 -- # return 0 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@515 -- # '[' -n 4155406 ']' 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@516 -- # killprocess 4155406 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@950 -- # '[' -z 4155406 ']' 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@954 -- # kill -0 4155406 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@955 -- # uname 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4155406 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4155406' 00:12:39.688 killing process with pid 4155406 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@969 -- # kill 4155406 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@974 -- # wait 4155406 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@297 -- # iptr 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@789 -- # iptables-save 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@789 -- # iptables-restore 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:39.688 17:33:47 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:42.231 00:12:42.231 real 0m38.120s 00:12:42.231 user 1m53.724s 00:12:42.231 sys 0m7.971s 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:12:42.231 ************************************ 00:12:42.231 END TEST nvmf_rpc 00:12:42.231 ************************************ 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@23 -- # run_test nvmf_invalid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:12:42.231 ************************************ 00:12:42.231 START TEST nvmf_invalid 00:12:42.231 ************************************ 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:12:42.231 * Looking for test storage... 00:12:42.231 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1691 -- # lcov --version 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # IFS=.-: 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # read -ra ver1 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # IFS=.-: 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # read -ra ver2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@338 -- # local 'op=<' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@340 -- # ver1_l=2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@341 -- # ver2_l=1 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@344 -- # case "$op" in 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@345 -- # : 1 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # decimal 1 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=1 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 1 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # ver1[v]=1 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # decimal 2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # ver2[v]=2 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # return 0 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:12:42.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:42.231 --rc genhtml_branch_coverage=1 00:12:42.231 --rc genhtml_function_coverage=1 00:12:42.231 --rc genhtml_legend=1 00:12:42.231 --rc geninfo_all_blocks=1 00:12:42.231 --rc geninfo_unexecuted_blocks=1 00:12:42.231 00:12:42.231 ' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:12:42.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:42.231 --rc genhtml_branch_coverage=1 00:12:42.231 --rc genhtml_function_coverage=1 00:12:42.231 --rc genhtml_legend=1 00:12:42.231 --rc geninfo_all_blocks=1 00:12:42.231 --rc geninfo_unexecuted_blocks=1 00:12:42.231 00:12:42.231 ' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:12:42.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:42.231 --rc genhtml_branch_coverage=1 00:12:42.231 --rc genhtml_function_coverage=1 00:12:42.231 --rc genhtml_legend=1 00:12:42.231 --rc geninfo_all_blocks=1 00:12:42.231 --rc geninfo_unexecuted_blocks=1 00:12:42.231 00:12:42.231 ' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:12:42.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:42.231 --rc genhtml_branch_coverage=1 00:12:42.231 --rc genhtml_function_coverage=1 00:12:42.231 --rc genhtml_legend=1 00:12:42.231 --rc geninfo_all_blocks=1 00:12:42.231 --rc geninfo_unexecuted_blocks=1 00:12:42.231 00:12:42.231 ' 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:42.231 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # uname -s 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@15 -- # shopt -s extglob 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@5 -- # export PATH 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@51 -- # : 0 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:42.232 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@11 -- # multi_target_rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@12 -- # rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@14 -- # target=foobar 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@16 -- # RANDOM=0 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@34 -- # nvmftestinit 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@474 -- # prepare_net_devs 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@436 -- # local -g is_hw=no 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@438 -- # remove_spdk_ns 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@309 -- # xtrace_disable 00:12:42.232 17:33:49 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # pci_devs=() 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # net_devs=() 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # e810=() 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # local -ga e810 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # x722=() 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # local -ga x722 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # mlx=() 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # local -ga mlx 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:12:50.382 Found 0000:31:00.0 (0x8086 - 0x159b) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:12:50.382 Found 0000:31:00.1 (0x8086 - 0x159b) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:12:50.382 Found net devices under 0000:31:00.0: cvl_0_0 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:12:50.382 Found net devices under 0000:31:00.1: cvl_0_1 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # is_hw=yes 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:50.382 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:50.382 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.571 ms 00:12:50.382 00:12:50.382 --- 10.0.0.2 ping statistics --- 00:12:50.382 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:50.382 rtt min/avg/max/mdev = 0.571/0.571/0.571/0.000 ms 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:50.382 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:50.382 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.278 ms 00:12:50.382 00:12:50.382 --- 10.0.0.1 ping statistics --- 00:12:50.382 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:50.382 rtt min/avg/max/mdev = 0.278/0.278/0.278/0.000 ms 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@448 -- # return 0 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:12:50.382 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@35 -- # nvmfappstart -m 0xF 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@724 -- # xtrace_disable 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@507 -- # nvmfpid=4165199 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@508 -- # waitforlisten 4165199 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@831 -- # '[' -z 4165199 ']' 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:50.383 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:50.383 17:33:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:12:50.383 [2024-10-17 17:33:57.649667] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:12:50.383 [2024-10-17 17:33:57.649750] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:50.383 [2024-10-17 17:33:57.738756] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:50.383 [2024-10-17 17:33:57.792610] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:50.383 [2024-10-17 17:33:57.792663] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:50.383 [2024-10-17 17:33:57.792671] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:50.383 [2024-10-17 17:33:57.792679] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:50.383 [2024-10-17 17:33:57.792685] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:50.383 [2024-10-17 17:33:57.794772] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:50.383 [2024-10-17 17:33:57.794936] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:12:50.383 [2024-10-17 17:33:57.795071] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:12:50.383 [2024-10-17 17:33:57.795072] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@864 -- # return 0 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@730 -- # xtrace_disable 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@37 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:12:50.644 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -t foobar nqn.2016-06.io.spdk:cnode32334 00:12:50.905 [2024-10-17 17:33:58.691705] nvmf_rpc.c: 396:rpc_nvmf_create_subsystem: *ERROR*: Unable to find target foobar 00:12:50.905 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # out='request: 00:12:50.905 { 00:12:50.905 "nqn": "nqn.2016-06.io.spdk:cnode32334", 00:12:50.905 "tgt_name": "foobar", 00:12:50.905 "method": "nvmf_create_subsystem", 00:12:50.905 "req_id": 1 00:12:50.905 } 00:12:50.905 Got JSON-RPC error response 00:12:50.905 response: 00:12:50.905 { 00:12:50.905 "code": -32603, 00:12:50.905 "message": "Unable to find target foobar" 00:12:50.905 }' 00:12:50.905 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@41 -- # [[ request: 00:12:50.905 { 00:12:50.905 "nqn": "nqn.2016-06.io.spdk:cnode32334", 00:12:50.905 "tgt_name": "foobar", 00:12:50.905 "method": "nvmf_create_subsystem", 00:12:50.905 "req_id": 1 00:12:50.905 } 00:12:50.905 Got JSON-RPC error response 00:12:50.905 response: 00:12:50.905 { 00:12:50.905 "code": -32603, 00:12:50.905 "message": "Unable to find target foobar" 00:12:50.905 } == *\U\n\a\b\l\e\ \t\o\ \f\i\n\d\ \t\a\r\g\e\t* ]] 00:12:50.905 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # echo -e '\x1f' 00:12:50.905 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s $'SPDKISFASTANDAWESOME\037' nqn.2016-06.io.spdk:cnode3497 00:12:51.165 [2024-10-17 17:33:58.900611] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode3497: invalid serial number 'SPDKISFASTANDAWESOME' 00:12:51.165 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # out='request: 00:12:51.165 { 00:12:51.165 "nqn": "nqn.2016-06.io.spdk:cnode3497", 00:12:51.165 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:12:51.165 "method": "nvmf_create_subsystem", 00:12:51.165 "req_id": 1 00:12:51.165 } 00:12:51.165 Got JSON-RPC error response 00:12:51.165 response: 00:12:51.165 { 00:12:51.165 "code": -32602, 00:12:51.165 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:12:51.165 }' 00:12:51.165 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@46 -- # [[ request: 00:12:51.165 { 00:12:51.165 "nqn": "nqn.2016-06.io.spdk:cnode3497", 00:12:51.165 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:12:51.165 "method": "nvmf_create_subsystem", 00:12:51.165 "req_id": 1 00:12:51.165 } 00:12:51.165 Got JSON-RPC error response 00:12:51.165 response: 00:12:51.165 { 00:12:51.165 "code": -32602, 00:12:51.166 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:12:51.166 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:12:51.166 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # echo -e '\x1f' 00:12:51.166 17:33:58 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d $'SPDK_Controller\037' nqn.2016-06.io.spdk:cnode30186 00:12:51.428 [2024-10-17 17:33:59.109361] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode30186: invalid model number 'SPDK_Controller' 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # out='request: 00:12:51.428 { 00:12:51.428 "nqn": "nqn.2016-06.io.spdk:cnode30186", 00:12:51.428 "model_number": "SPDK_Controller\u001f", 00:12:51.428 "method": "nvmf_create_subsystem", 00:12:51.428 "req_id": 1 00:12:51.428 } 00:12:51.428 Got JSON-RPC error response 00:12:51.428 response: 00:12:51.428 { 00:12:51.428 "code": -32602, 00:12:51.428 "message": "Invalid MN SPDK_Controller\u001f" 00:12:51.428 }' 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@51 -- # [[ request: 00:12:51.428 { 00:12:51.428 "nqn": "nqn.2016-06.io.spdk:cnode30186", 00:12:51.428 "model_number": "SPDK_Controller\u001f", 00:12:51.428 "method": "nvmf_create_subsystem", 00:12:51.428 "req_id": 1 00:12:51.428 } 00:12:51.428 Got JSON-RPC error response 00:12:51.428 response: 00:12:51.428 { 00:12:51.428 "code": -32602, 00:12:51.428 "message": "Invalid MN SPDK_Controller\u001f" 00:12:51.428 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # gen_random_s 21 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@19 -- # local length=21 ll 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 104 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x68' 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=h 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 107 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6b' 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=k 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 117 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x75' 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=u 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 88 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x58' 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=X 00:12:51.428 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 57 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x39' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=9 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 90 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5a' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Z 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 48 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x30' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=0 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 50 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x32' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=2 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 40 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x28' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='(' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 86 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x56' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=V 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 126 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7e' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='~' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 80 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x50' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=P 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 47 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2f' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=/ 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 70 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x46' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=F 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 63 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3f' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='?' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 74 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4a' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=J 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 64 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x40' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=@ 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 58 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3a' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=: 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 64 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x40' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=@ 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 72 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x48' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=H 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 119 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x77' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=w 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@28 -- # [[ h == \- ]] 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@31 -- # echo 'hkuX9Z02(V~P/F?J@:@Hw' 00:12:51.429 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s 'hkuX9Z02(V~P/F?J@:@Hw' nqn.2016-06.io.spdk:cnode17279 00:12:51.691 [2024-10-17 17:33:59.498875] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode17279: invalid serial number 'hkuX9Z02(V~P/F?J@:@Hw' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # out='request: 00:12:51.691 { 00:12:51.691 "nqn": "nqn.2016-06.io.spdk:cnode17279", 00:12:51.691 "serial_number": "hkuX9Z02(V~P/F?J@:@Hw", 00:12:51.691 "method": "nvmf_create_subsystem", 00:12:51.691 "req_id": 1 00:12:51.691 } 00:12:51.691 Got JSON-RPC error response 00:12:51.691 response: 00:12:51.691 { 00:12:51.691 "code": -32602, 00:12:51.691 "message": "Invalid SN hkuX9Z02(V~P/F?J@:@Hw" 00:12:51.691 }' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@55 -- # [[ request: 00:12:51.691 { 00:12:51.691 "nqn": "nqn.2016-06.io.spdk:cnode17279", 00:12:51.691 "serial_number": "hkuX9Z02(V~P/F?J@:@Hw", 00:12:51.691 "method": "nvmf_create_subsystem", 00:12:51.691 "req_id": 1 00:12:51.691 } 00:12:51.691 Got JSON-RPC error response 00:12:51.691 response: 00:12:51.691 { 00:12:51.691 "code": -32602, 00:12:51.691 "message": "Invalid SN hkuX9Z02(V~P/F?J@:@Hw" 00:12:51.691 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # gen_random_s 41 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@19 -- # local length=41 ll 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 114 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x72' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=r 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 85 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x55' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=U 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 40 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x28' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='(' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 45 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2d' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=- 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 103 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x67' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=g 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 66 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x42' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=B 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 90 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5a' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Z 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 86 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x56' 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=V 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.691 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 124 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7c' 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='|' 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 82 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x52' 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=R 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 98 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x62' 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=b 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 108 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6c' 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=l 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 77 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4d' 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=M 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.953 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 77 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4d' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=M 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 35 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x23' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='#' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 71 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x47' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=G 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 93 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5d' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=']' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 123 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7b' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='{' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 43 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2b' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=+ 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 56 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x38' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=8 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 119 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x77' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=w 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 33 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x21' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='!' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 122 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7a' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=z 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 65 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x41' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=A 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 38 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x26' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='&' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 45 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2d' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=- 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 93 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5d' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=']' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 90 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5a' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Z 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 64 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x40' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=@ 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 89 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x59' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Y 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 53 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x35' 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=5 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.954 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 69 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x45' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=E 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 86 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x56' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=V 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 50 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x32' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=2 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 60 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3c' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='<' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 47 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2f' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=/ 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 120 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x78' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=x 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 62 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3e' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='>' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 72 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x48' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=H 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 110 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6e' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=n 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 69 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x45' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=E 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@28 -- # [[ r == \- ]] 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@31 -- # echo 'rU(-gBZV|RblMM#G]{+8w!zA&-]Z@Y5EV2HnE' 00:12:51.955 17:33:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d 'rU(-gBZV|RblMM#G]{+8w!zA&-]Z@Y5EV2HnE' nqn.2016-06.io.spdk:cnode870 00:12:52.216 [2024-10-17 17:34:00.016742] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode870: invalid model number 'rU(-gBZV|RblMM#G]{+8w!zA&-]Z@Y5EV2HnE' 00:12:52.216 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # out='request: 00:12:52.216 { 00:12:52.216 "nqn": "nqn.2016-06.io.spdk:cnode870", 00:12:52.216 "model_number": "rU(-gBZV|RblMM#G]{+8w!zA&-]Z@Y5EV2HnE", 00:12:52.216 "method": "nvmf_create_subsystem", 00:12:52.216 "req_id": 1 00:12:52.216 } 00:12:52.216 Got JSON-RPC error response 00:12:52.216 response: 00:12:52.216 { 00:12:52.216 "code": -32602, 00:12:52.216 "message": "Invalid MN rU(-gBZV|RblMM#G]{+8w!zA&-]Z@Y5EV2HnE" 00:12:52.216 }' 00:12:52.216 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@59 -- # [[ request: 00:12:52.216 { 00:12:52.216 "nqn": "nqn.2016-06.io.spdk:cnode870", 00:12:52.216 "model_number": "rU(-gBZV|RblMM#G]{+8w!zA&-]Z@Y5EV2HnE", 00:12:52.216 "method": "nvmf_create_subsystem", 00:12:52.216 "req_id": 1 00:12:52.216 } 00:12:52.216 Got JSON-RPC error response 00:12:52.216 response: 00:12:52.216 { 00:12:52.216 "code": -32602, 00:12:52.216 "message": "Invalid MN rU(-gBZV|RblMM#G]{+8w!zA&-]Z@Y5EV2HnE" 00:12:52.216 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:12:52.216 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport --trtype tcp 00:12:52.477 [2024-10-17 17:34:00.197421] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:52.477 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode -s SPDK001 -a 00:12:52.737 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@64 -- # [[ tcp == \T\C\P ]] 00:12:52.737 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # echo '' 00:12:52.737 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # head -n 1 00:12:52.737 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # IP= 00:12:52.737 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode -t tcp -a '' -s 4421 00:12:52.737 [2024-10-17 17:34:00.582627] nvmf_rpc.c: 783:nvmf_rpc_listen_paused: *ERROR*: Unable to remove listener, rc -2 00:12:52.737 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # out='request: 00:12:52.737 { 00:12:52.737 "nqn": "nqn.2016-06.io.spdk:cnode", 00:12:52.737 "listen_address": { 00:12:52.737 "trtype": "tcp", 00:12:52.737 "traddr": "", 00:12:52.737 "trsvcid": "4421" 00:12:52.737 }, 00:12:52.737 "method": "nvmf_subsystem_remove_listener", 00:12:52.737 "req_id": 1 00:12:52.737 } 00:12:52.737 Got JSON-RPC error response 00:12:52.737 response: 00:12:52.737 { 00:12:52.737 "code": -32602, 00:12:52.737 "message": "Invalid parameters" 00:12:52.737 }' 00:12:52.737 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@70 -- # [[ request: 00:12:52.737 { 00:12:52.737 "nqn": "nqn.2016-06.io.spdk:cnode", 00:12:52.737 "listen_address": { 00:12:52.737 "trtype": "tcp", 00:12:52.737 "traddr": "", 00:12:52.737 "trsvcid": "4421" 00:12:52.737 }, 00:12:52.737 "method": "nvmf_subsystem_remove_listener", 00:12:52.737 "req_id": 1 00:12:52.737 } 00:12:52.737 Got JSON-RPC error response 00:12:52.737 response: 00:12:52.738 { 00:12:52.738 "code": -32602, 00:12:52.738 "message": "Invalid parameters" 00:12:52.738 } != *\U\n\a\b\l\e\ \t\o\ \s\t\o\p\ \l\i\s\t\e\n\e\r\.* ]] 00:12:52.738 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode26215 -i 0 00:12:52.998 [2024-10-17 17:34:00.775187] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode26215: invalid cntlid range [0-65519] 00:12:52.998 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # out='request: 00:12:52.998 { 00:12:52.998 "nqn": "nqn.2016-06.io.spdk:cnode26215", 00:12:52.998 "min_cntlid": 0, 00:12:52.998 "method": "nvmf_create_subsystem", 00:12:52.998 "req_id": 1 00:12:52.998 } 00:12:52.998 Got JSON-RPC error response 00:12:52.998 response: 00:12:52.998 { 00:12:52.998 "code": -32602, 00:12:52.998 "message": "Invalid cntlid range [0-65519]" 00:12:52.998 }' 00:12:52.998 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@74 -- # [[ request: 00:12:52.998 { 00:12:52.998 "nqn": "nqn.2016-06.io.spdk:cnode26215", 00:12:52.998 "min_cntlid": 0, 00:12:52.998 "method": "nvmf_create_subsystem", 00:12:52.998 "req_id": 1 00:12:52.998 } 00:12:52.998 Got JSON-RPC error response 00:12:52.998 response: 00:12:52.998 { 00:12:52.998 "code": -32602, 00:12:52.998 "message": "Invalid cntlid range [0-65519]" 00:12:52.998 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:12:52.998 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode9221 -i 65520 00:12:53.260 [2024-10-17 17:34:00.963836] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode9221: invalid cntlid range [65520-65519] 00:12:53.260 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # out='request: 00:12:53.260 { 00:12:53.260 "nqn": "nqn.2016-06.io.spdk:cnode9221", 00:12:53.260 "min_cntlid": 65520, 00:12:53.260 "method": "nvmf_create_subsystem", 00:12:53.260 "req_id": 1 00:12:53.260 } 00:12:53.260 Got JSON-RPC error response 00:12:53.260 response: 00:12:53.260 { 00:12:53.260 "code": -32602, 00:12:53.260 "message": "Invalid cntlid range [65520-65519]" 00:12:53.260 }' 00:12:53.260 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@76 -- # [[ request: 00:12:53.260 { 00:12:53.260 "nqn": "nqn.2016-06.io.spdk:cnode9221", 00:12:53.260 "min_cntlid": 65520, 00:12:53.260 "method": "nvmf_create_subsystem", 00:12:53.260 "req_id": 1 00:12:53.260 } 00:12:53.260 Got JSON-RPC error response 00:12:53.260 response: 00:12:53.260 { 00:12:53.260 "code": -32602, 00:12:53.260 "message": "Invalid cntlid range [65520-65519]" 00:12:53.260 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:12:53.260 17:34:00 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode14522 -I 0 00:12:53.260 [2024-10-17 17:34:01.152394] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode14522: invalid cntlid range [1-0] 00:12:53.520 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # out='request: 00:12:53.520 { 00:12:53.520 "nqn": "nqn.2016-06.io.spdk:cnode14522", 00:12:53.520 "max_cntlid": 0, 00:12:53.520 "method": "nvmf_create_subsystem", 00:12:53.520 "req_id": 1 00:12:53.520 } 00:12:53.520 Got JSON-RPC error response 00:12:53.520 response: 00:12:53.520 { 00:12:53.520 "code": -32602, 00:12:53.520 "message": "Invalid cntlid range [1-0]" 00:12:53.520 }' 00:12:53.520 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@78 -- # [[ request: 00:12:53.520 { 00:12:53.520 "nqn": "nqn.2016-06.io.spdk:cnode14522", 00:12:53.520 "max_cntlid": 0, 00:12:53.520 "method": "nvmf_create_subsystem", 00:12:53.520 "req_id": 1 00:12:53.520 } 00:12:53.520 Got JSON-RPC error response 00:12:53.520 response: 00:12:53.520 { 00:12:53.520 "code": -32602, 00:12:53.520 "message": "Invalid cntlid range [1-0]" 00:12:53.520 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:12:53.520 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode18975 -I 65520 00:12:53.520 [2024-10-17 17:34:01.340984] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode18975: invalid cntlid range [1-65520] 00:12:53.520 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # out='request: 00:12:53.520 { 00:12:53.520 "nqn": "nqn.2016-06.io.spdk:cnode18975", 00:12:53.520 "max_cntlid": 65520, 00:12:53.520 "method": "nvmf_create_subsystem", 00:12:53.520 "req_id": 1 00:12:53.520 } 00:12:53.520 Got JSON-RPC error response 00:12:53.520 response: 00:12:53.520 { 00:12:53.520 "code": -32602, 00:12:53.520 "message": "Invalid cntlid range [1-65520]" 00:12:53.520 }' 00:12:53.520 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@80 -- # [[ request: 00:12:53.520 { 00:12:53.520 "nqn": "nqn.2016-06.io.spdk:cnode18975", 00:12:53.520 "max_cntlid": 65520, 00:12:53.521 "method": "nvmf_create_subsystem", 00:12:53.521 "req_id": 1 00:12:53.521 } 00:12:53.521 Got JSON-RPC error response 00:12:53.521 response: 00:12:53.521 { 00:12:53.521 "code": -32602, 00:12:53.521 "message": "Invalid cntlid range [1-65520]" 00:12:53.521 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:12:53.521 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode14577 -i 6 -I 5 00:12:53.781 [2024-10-17 17:34:01.529582] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode14577: invalid cntlid range [6-5] 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # out='request: 00:12:53.781 { 00:12:53.781 "nqn": "nqn.2016-06.io.spdk:cnode14577", 00:12:53.781 "min_cntlid": 6, 00:12:53.781 "max_cntlid": 5, 00:12:53.781 "method": "nvmf_create_subsystem", 00:12:53.781 "req_id": 1 00:12:53.781 } 00:12:53.781 Got JSON-RPC error response 00:12:53.781 response: 00:12:53.781 { 00:12:53.781 "code": -32602, 00:12:53.781 "message": "Invalid cntlid range [6-5]" 00:12:53.781 }' 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@84 -- # [[ request: 00:12:53.781 { 00:12:53.781 "nqn": "nqn.2016-06.io.spdk:cnode14577", 00:12:53.781 "min_cntlid": 6, 00:12:53.781 "max_cntlid": 5, 00:12:53.781 "method": "nvmf_create_subsystem", 00:12:53.781 "req_id": 1 00:12:53.781 } 00:12:53.781 Got JSON-RPC error response 00:12:53.781 response: 00:12:53.781 { 00:12:53.781 "code": -32602, 00:12:53.781 "message": "Invalid cntlid range [6-5]" 00:12:53.781 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target --name foobar 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # out='request: 00:12:53.781 { 00:12:53.781 "name": "foobar", 00:12:53.781 "method": "nvmf_delete_target", 00:12:53.781 "req_id": 1 00:12:53.781 } 00:12:53.781 Got JSON-RPC error response 00:12:53.781 response: 00:12:53.781 { 00:12:53.781 "code": -32602, 00:12:53.781 "message": "The specified target doesn'\''t exist, cannot delete it." 00:12:53.781 }' 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@88 -- # [[ request: 00:12:53.781 { 00:12:53.781 "name": "foobar", 00:12:53.781 "method": "nvmf_delete_target", 00:12:53.781 "req_id": 1 00:12:53.781 } 00:12:53.781 Got JSON-RPC error response 00:12:53.781 response: 00:12:53.781 { 00:12:53.781 "code": -32602, 00:12:53.781 "message": "The specified target doesn't exist, cannot delete it." 00:12:53.781 } == *\T\h\e\ \s\p\e\c\i\f\i\e\d\ \t\a\r\g\e\t\ \d\o\e\s\n\'\t\ \e\x\i\s\t\,\ \c\a\n\n\o\t\ \d\e\l\e\t\e\ \i\t\.* ]] 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@90 -- # trap - SIGINT SIGTERM EXIT 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@91 -- # nvmftestfini 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@514 -- # nvmfcleanup 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@121 -- # sync 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@124 -- # set +e 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:53.781 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:53.781 rmmod nvme_tcp 00:12:53.781 rmmod nvme_fabrics 00:12:54.042 rmmod nvme_keyring 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@128 -- # set -e 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@129 -- # return 0 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@515 -- # '[' -n 4165199 ']' 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@516 -- # killprocess 4165199 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@950 -- # '[' -z 4165199 ']' 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@954 -- # kill -0 4165199 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@955 -- # uname 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4165199 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4165199' 00:12:54.042 killing process with pid 4165199 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@969 -- # kill 4165199 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@974 -- # wait 4165199 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@297 -- # iptr 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@789 -- # iptables-save 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@789 -- # iptables-restore 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:54.042 17:34:01 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:56.587 17:34:03 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:56.587 00:12:56.587 real 0m14.247s 00:12:56.587 user 0m21.123s 00:12:56.587 sys 0m6.765s 00:12:56.587 17:34:03 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:56.587 17:34:03 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:12:56.587 ************************************ 00:12:56.587 END TEST nvmf_invalid 00:12:56.587 ************************************ 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@24 -- # run_test nvmf_connect_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:12:56.587 ************************************ 00:12:56.587 START TEST nvmf_connect_stress 00:12:56.587 ************************************ 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:12:56.587 * Looking for test storage... 00:12:56.587 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1691 -- # lcov --version 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # IFS=.-: 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # read -ra ver1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # IFS=.-: 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # read -ra ver2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@338 -- # local 'op=<' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@340 -- # ver1_l=2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@341 -- # ver2_l=1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@344 -- # case "$op" in 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@345 -- # : 1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # decimal 1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # decimal 2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # return 0 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:12:56.587 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.587 --rc genhtml_branch_coverage=1 00:12:56.587 --rc genhtml_function_coverage=1 00:12:56.587 --rc genhtml_legend=1 00:12:56.587 --rc geninfo_all_blocks=1 00:12:56.587 --rc geninfo_unexecuted_blocks=1 00:12:56.587 00:12:56.587 ' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:12:56.587 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.587 --rc genhtml_branch_coverage=1 00:12:56.587 --rc genhtml_function_coverage=1 00:12:56.587 --rc genhtml_legend=1 00:12:56.587 --rc geninfo_all_blocks=1 00:12:56.587 --rc geninfo_unexecuted_blocks=1 00:12:56.587 00:12:56.587 ' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:12:56.587 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.587 --rc genhtml_branch_coverage=1 00:12:56.587 --rc genhtml_function_coverage=1 00:12:56.587 --rc genhtml_legend=1 00:12:56.587 --rc geninfo_all_blocks=1 00:12:56.587 --rc geninfo_unexecuted_blocks=1 00:12:56.587 00:12:56.587 ' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:12:56.587 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.587 --rc genhtml_branch_coverage=1 00:12:56.587 --rc genhtml_function_coverage=1 00:12:56.587 --rc genhtml_legend=1 00:12:56.587 --rc geninfo_all_blocks=1 00:12:56.587 --rc geninfo_unexecuted_blocks=1 00:12:56.587 00:12:56.587 ' 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # uname -s 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:56.587 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@5 -- # export PATH 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@51 -- # : 0 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:56.588 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@12 -- # nvmftestinit 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@474 -- # prepare_net_devs 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@436 -- # local -g is_hw=no 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@438 -- # remove_spdk_ns 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:12:56.588 17:34:04 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # net_devs=() 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # e810=() 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # local -ga e810 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # x722=() 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # local -ga x722 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # mlx=() 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:04.730 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:13:04.731 Found 0000:31:00.0 (0x8086 - 0x159b) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:13:04.731 Found 0000:31:00.1 (0x8086 - 0x159b) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:13:04.731 Found net devices under 0000:31:00.0: cvl_0_0 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:13:04.731 Found net devices under 0000:31:00.1: cvl_0_1 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # is_hw=yes 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:04.731 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:04.731 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.547 ms 00:13:04.731 00:13:04.731 --- 10.0.0.2 ping statistics --- 00:13:04.731 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:04.731 rtt min/avg/max/mdev = 0.547/0.547/0.547/0.000 ms 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:04.731 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:04.731 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.296 ms 00:13:04.731 00:13:04.731 --- 10.0.0.1 ping statistics --- 00:13:04.731 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:04.731 rtt min/avg/max/mdev = 0.296/0.296/0.296/0.000 ms 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@448 -- # return 0 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@13 -- # nvmfappstart -m 0xE 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@724 -- # xtrace_disable 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.731 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@507 -- # nvmfpid=4170426 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@508 -- # waitforlisten 4170426 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@831 -- # '[' -z 4170426 ']' 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:04.732 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:04.732 17:34:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.732 [2024-10-17 17:34:11.984277] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:13:04.732 [2024-10-17 17:34:11.984343] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:04.732 [2024-10-17 17:34:12.073720] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:13:04.732 [2024-10-17 17:34:12.125499] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:04.732 [2024-10-17 17:34:12.125551] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:04.732 [2024-10-17 17:34:12.125560] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:04.732 [2024-10-17 17:34:12.125567] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:04.732 [2024-10-17 17:34:12.125573] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:04.732 [2024-10-17 17:34:12.127503] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:04.732 [2024-10-17 17:34:12.127663] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:04.732 [2024-10-17 17:34:12.127665] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@864 -- # return 0 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@730 -- # xtrace_disable 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.993 [2024-10-17 17:34:12.862141] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.993 [2024-10-17 17:34:12.887739] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:04.993 NULL1 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@21 -- # PERF_PID=4170553 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@23 -- # rpcs=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/connect_stress/connect_stress -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -t 10 00:13:04.993 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@25 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # seq 1 20 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.255 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:05.516 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.516 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:05.516 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:05.516 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.516 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:05.777 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.777 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:05.777 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:05.777 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.777 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:06.348 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.348 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:06.348 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:06.348 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.348 17:34:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:06.608 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.608 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:06.608 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:06.608 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.608 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:06.868 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.868 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:06.868 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:06.868 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.868 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:07.128 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.128 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:07.128 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:07.128 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.128 17:34:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:07.388 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.388 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:07.388 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:07.388 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.388 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:07.958 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.958 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:07.958 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:07.958 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.958 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:08.219 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.219 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:08.219 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:08.219 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.219 17:34:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:08.479 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.479 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:08.479 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:08.479 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.479 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:08.739 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.739 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:08.739 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:08.739 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.739 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:08.999 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.999 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:09.258 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:09.258 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.258 17:34:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:09.526 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.526 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:09.526 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:09.526 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.526 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:09.789 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.789 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:09.789 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:09.789 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.789 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:10.049 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.049 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:10.049 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:10.049 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.049 17:34:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:10.309 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.309 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:10.309 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:10.309 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.309 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:10.878 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.878 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:10.878 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:10.878 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.878 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:11.178 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.178 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:11.178 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:11.178 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.178 17:34:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:11.438 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.438 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:11.438 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:11.438 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.438 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:11.698 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.698 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:11.698 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:11.698 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.698 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:11.957 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.957 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:11.957 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:11.957 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.957 17:34:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:12.526 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.526 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:12.526 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:12.526 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.526 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:12.786 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.786 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:12.786 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:12.786 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.786 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:13.045 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.045 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:13.045 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:13.045 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.045 17:34:20 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:13.305 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.305 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:13.305 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:13.305 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.305 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:13.565 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.565 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:13.565 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:13.565 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.565 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:14.134 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.134 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:14.134 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:14.134 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.134 17:34:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:14.394 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.394 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:14.394 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:14.394 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.394 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:14.655 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.655 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:14.655 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:14.655 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.655 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:14.915 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.915 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:14.915 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:13:14.915 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.915 17:34:22 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:15.175 Testing NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 4170553 00:13:15.435 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh: line 34: kill: (4170553) - No such process 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@38 -- # wait 4170553 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@39 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@43 -- # nvmftestfini 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@514 -- # nvmfcleanup 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@121 -- # sync 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@124 -- # set +e 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:15.435 rmmod nvme_tcp 00:13:15.435 rmmod nvme_fabrics 00:13:15.435 rmmod nvme_keyring 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@128 -- # set -e 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@129 -- # return 0 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@515 -- # '[' -n 4170426 ']' 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@516 -- # killprocess 4170426 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@950 -- # '[' -z 4170426 ']' 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@954 -- # kill -0 4170426 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@955 -- # uname 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4170426 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4170426' 00:13:15.435 killing process with pid 4170426 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@969 -- # kill 4170426 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@974 -- # wait 4170426 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@297 -- # iptr 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@789 -- # iptables-save 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@789 -- # iptables-restore 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:15.435 17:34:23 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:18.080 00:13:18.080 real 0m21.345s 00:13:18.080 user 0m42.068s 00:13:18.080 sys 0m9.445s 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:13:18.080 ************************************ 00:13:18.080 END TEST nvmf_connect_stress 00:13:18.080 ************************************ 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@25 -- # run_test nvmf_fused_ordering /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:13:18.080 ************************************ 00:13:18.080 START TEST nvmf_fused_ordering 00:13:18.080 ************************************ 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:13:18.080 * Looking for test storage... 00:13:18.080 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1691 -- # lcov --version 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # IFS=.-: 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # read -ra ver1 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # IFS=.-: 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # read -ra ver2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@338 -- # local 'op=<' 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@340 -- # ver1_l=2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@341 -- # ver2_l=1 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@344 -- # case "$op" in 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@345 -- # : 1 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # decimal 1 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=1 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 1 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # ver1[v]=1 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # decimal 2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # ver2[v]=2 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # return 0 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:13:18.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:18.080 --rc genhtml_branch_coverage=1 00:13:18.080 --rc genhtml_function_coverage=1 00:13:18.080 --rc genhtml_legend=1 00:13:18.080 --rc geninfo_all_blocks=1 00:13:18.080 --rc geninfo_unexecuted_blocks=1 00:13:18.080 00:13:18.080 ' 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:13:18.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:18.080 --rc genhtml_branch_coverage=1 00:13:18.080 --rc genhtml_function_coverage=1 00:13:18.080 --rc genhtml_legend=1 00:13:18.080 --rc geninfo_all_blocks=1 00:13:18.080 --rc geninfo_unexecuted_blocks=1 00:13:18.080 00:13:18.080 ' 00:13:18.080 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:13:18.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:18.080 --rc genhtml_branch_coverage=1 00:13:18.081 --rc genhtml_function_coverage=1 00:13:18.081 --rc genhtml_legend=1 00:13:18.081 --rc geninfo_all_blocks=1 00:13:18.081 --rc geninfo_unexecuted_blocks=1 00:13:18.081 00:13:18.081 ' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:13:18.081 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:18.081 --rc genhtml_branch_coverage=1 00:13:18.081 --rc genhtml_function_coverage=1 00:13:18.081 --rc genhtml_legend=1 00:13:18.081 --rc geninfo_all_blocks=1 00:13:18.081 --rc geninfo_unexecuted_blocks=1 00:13:18.081 00:13:18.081 ' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # uname -s 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@15 -- # shopt -s extglob 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@5 -- # export PATH 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@51 -- # : 0 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:18.081 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@12 -- # nvmftestinit 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@474 -- # prepare_net_devs 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@436 -- # local -g is_hw=no 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@438 -- # remove_spdk_ns 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@309 -- # xtrace_disable 00:13:18.081 17:34:25 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # pci_devs=() 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # net_devs=() 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # e810=() 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # local -ga e810 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # x722=() 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # local -ga x722 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # mlx=() 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # local -ga mlx 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:13:26.224 Found 0000:31:00.0 (0x8086 - 0x159b) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:13:26.224 Found 0000:31:00.1 (0x8086 - 0x159b) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:13:26.224 Found net devices under 0000:31:00.0: cvl_0_0 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:13:26.224 Found net devices under 0000:31:00.1: cvl_0_1 00:13:26.224 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # is_hw=yes 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:26.225 17:34:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:26.225 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:26.225 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.679 ms 00:13:26.225 00:13:26.225 --- 10.0.0.2 ping statistics --- 00:13:26.225 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:26.225 rtt min/avg/max/mdev = 0.679/0.679/0.679/0.000 ms 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:26.225 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:26.225 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.321 ms 00:13:26.225 00:13:26.225 --- 10.0.0.1 ping statistics --- 00:13:26.225 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:26.225 rtt min/avg/max/mdev = 0.321/0.321/0.321/0.000 ms 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@448 -- # return 0 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@13 -- # nvmfappstart -m 0x2 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@724 -- # xtrace_disable 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@507 -- # nvmfpid=4176938 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@508 -- # waitforlisten 4176938 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@831 -- # '[' -z 4176938 ']' 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:26.225 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:26.225 17:34:33 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.225 [2024-10-17 17:34:33.433239] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:13:26.225 [2024-10-17 17:34:33.433329] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:26.225 [2024-10-17 17:34:33.525381] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:26.225 [2024-10-17 17:34:33.575911] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:26.225 [2024-10-17 17:34:33.575962] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:26.225 [2024-10-17 17:34:33.575971] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:26.225 [2024-10-17 17:34:33.575978] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:26.225 [2024-10-17 17:34:33.575984] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:26.225 [2024-10-17 17:34:33.576817] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@864 -- # return 0 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@730 -- # xtrace_disable 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.486 [2024-10-17 17:34:34.302334] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.486 [2024-10-17 17:34:34.326652] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.486 NULL1 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@19 -- # rpc_cmd bdev_wait_for_examine 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.486 17:34:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/fused_ordering/fused_ordering -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:13:26.486 [2024-10-17 17:34:34.395663] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:13:26.486 [2024-10-17 17:34:34.395715] [ DPDK EAL parameters: fused_ordering --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4176980 ] 00:13:27.060 Attached to nqn.2016-06.io.spdk:cnode1 00:13:27.060 Namespace ID: 1 size: 1GB 00:13:27.060 fused_ordering(0) 00:13:27.060 fused_ordering(1) 00:13:27.060 fused_ordering(2) 00:13:27.060 fused_ordering(3) 00:13:27.060 fused_ordering(4) 00:13:27.060 fused_ordering(5) 00:13:27.060 fused_ordering(6) 00:13:27.060 fused_ordering(7) 00:13:27.060 fused_ordering(8) 00:13:27.060 fused_ordering(9) 00:13:27.060 fused_ordering(10) 00:13:27.060 fused_ordering(11) 00:13:27.060 fused_ordering(12) 00:13:27.060 fused_ordering(13) 00:13:27.060 fused_ordering(14) 00:13:27.060 fused_ordering(15) 00:13:27.060 fused_ordering(16) 00:13:27.060 fused_ordering(17) 00:13:27.060 fused_ordering(18) 00:13:27.060 fused_ordering(19) 00:13:27.060 fused_ordering(20) 00:13:27.060 fused_ordering(21) 00:13:27.060 fused_ordering(22) 00:13:27.060 fused_ordering(23) 00:13:27.060 fused_ordering(24) 00:13:27.060 fused_ordering(25) 00:13:27.060 fused_ordering(26) 00:13:27.060 fused_ordering(27) 00:13:27.060 fused_ordering(28) 00:13:27.060 fused_ordering(29) 00:13:27.060 fused_ordering(30) 00:13:27.060 fused_ordering(31) 00:13:27.060 fused_ordering(32) 00:13:27.060 fused_ordering(33) 00:13:27.060 fused_ordering(34) 00:13:27.060 fused_ordering(35) 00:13:27.060 fused_ordering(36) 00:13:27.060 fused_ordering(37) 00:13:27.060 fused_ordering(38) 00:13:27.060 fused_ordering(39) 00:13:27.060 fused_ordering(40) 00:13:27.060 fused_ordering(41) 00:13:27.060 fused_ordering(42) 00:13:27.060 fused_ordering(43) 00:13:27.060 fused_ordering(44) 00:13:27.060 fused_ordering(45) 00:13:27.060 fused_ordering(46) 00:13:27.060 fused_ordering(47) 00:13:27.060 fused_ordering(48) 00:13:27.060 fused_ordering(49) 00:13:27.060 fused_ordering(50) 00:13:27.060 fused_ordering(51) 00:13:27.060 fused_ordering(52) 00:13:27.060 fused_ordering(53) 00:13:27.060 fused_ordering(54) 00:13:27.060 fused_ordering(55) 00:13:27.060 fused_ordering(56) 00:13:27.060 fused_ordering(57) 00:13:27.060 fused_ordering(58) 00:13:27.060 fused_ordering(59) 00:13:27.060 fused_ordering(60) 00:13:27.060 fused_ordering(61) 00:13:27.060 fused_ordering(62) 00:13:27.060 fused_ordering(63) 00:13:27.060 fused_ordering(64) 00:13:27.060 fused_ordering(65) 00:13:27.060 fused_ordering(66) 00:13:27.060 fused_ordering(67) 00:13:27.060 fused_ordering(68) 00:13:27.060 fused_ordering(69) 00:13:27.060 fused_ordering(70) 00:13:27.060 fused_ordering(71) 00:13:27.060 fused_ordering(72) 00:13:27.060 fused_ordering(73) 00:13:27.060 fused_ordering(74) 00:13:27.060 fused_ordering(75) 00:13:27.060 fused_ordering(76) 00:13:27.060 fused_ordering(77) 00:13:27.060 fused_ordering(78) 00:13:27.060 fused_ordering(79) 00:13:27.060 fused_ordering(80) 00:13:27.060 fused_ordering(81) 00:13:27.060 fused_ordering(82) 00:13:27.060 fused_ordering(83) 00:13:27.060 fused_ordering(84) 00:13:27.060 fused_ordering(85) 00:13:27.060 fused_ordering(86) 00:13:27.060 fused_ordering(87) 00:13:27.060 fused_ordering(88) 00:13:27.060 fused_ordering(89) 00:13:27.060 fused_ordering(90) 00:13:27.060 fused_ordering(91) 00:13:27.060 fused_ordering(92) 00:13:27.060 fused_ordering(93) 00:13:27.060 fused_ordering(94) 00:13:27.060 fused_ordering(95) 00:13:27.060 fused_ordering(96) 00:13:27.060 fused_ordering(97) 00:13:27.060 fused_ordering(98) 00:13:27.060 fused_ordering(99) 00:13:27.060 fused_ordering(100) 00:13:27.060 fused_ordering(101) 00:13:27.060 fused_ordering(102) 00:13:27.060 fused_ordering(103) 00:13:27.060 fused_ordering(104) 00:13:27.060 fused_ordering(105) 00:13:27.060 fused_ordering(106) 00:13:27.060 fused_ordering(107) 00:13:27.060 fused_ordering(108) 00:13:27.060 fused_ordering(109) 00:13:27.060 fused_ordering(110) 00:13:27.060 fused_ordering(111) 00:13:27.060 fused_ordering(112) 00:13:27.060 fused_ordering(113) 00:13:27.060 fused_ordering(114) 00:13:27.060 fused_ordering(115) 00:13:27.060 fused_ordering(116) 00:13:27.060 fused_ordering(117) 00:13:27.060 fused_ordering(118) 00:13:27.060 fused_ordering(119) 00:13:27.060 fused_ordering(120) 00:13:27.060 fused_ordering(121) 00:13:27.060 fused_ordering(122) 00:13:27.060 fused_ordering(123) 00:13:27.060 fused_ordering(124) 00:13:27.060 fused_ordering(125) 00:13:27.060 fused_ordering(126) 00:13:27.060 fused_ordering(127) 00:13:27.060 fused_ordering(128) 00:13:27.060 fused_ordering(129) 00:13:27.060 fused_ordering(130) 00:13:27.060 fused_ordering(131) 00:13:27.060 fused_ordering(132) 00:13:27.060 fused_ordering(133) 00:13:27.060 fused_ordering(134) 00:13:27.060 fused_ordering(135) 00:13:27.060 fused_ordering(136) 00:13:27.060 fused_ordering(137) 00:13:27.060 fused_ordering(138) 00:13:27.060 fused_ordering(139) 00:13:27.060 fused_ordering(140) 00:13:27.060 fused_ordering(141) 00:13:27.060 fused_ordering(142) 00:13:27.060 fused_ordering(143) 00:13:27.060 fused_ordering(144) 00:13:27.060 fused_ordering(145) 00:13:27.060 fused_ordering(146) 00:13:27.060 fused_ordering(147) 00:13:27.060 fused_ordering(148) 00:13:27.060 fused_ordering(149) 00:13:27.060 fused_ordering(150) 00:13:27.060 fused_ordering(151) 00:13:27.060 fused_ordering(152) 00:13:27.060 fused_ordering(153) 00:13:27.060 fused_ordering(154) 00:13:27.060 fused_ordering(155) 00:13:27.060 fused_ordering(156) 00:13:27.060 fused_ordering(157) 00:13:27.060 fused_ordering(158) 00:13:27.060 fused_ordering(159) 00:13:27.060 fused_ordering(160) 00:13:27.060 fused_ordering(161) 00:13:27.060 fused_ordering(162) 00:13:27.060 fused_ordering(163) 00:13:27.060 fused_ordering(164) 00:13:27.060 fused_ordering(165) 00:13:27.060 fused_ordering(166) 00:13:27.060 fused_ordering(167) 00:13:27.060 fused_ordering(168) 00:13:27.060 fused_ordering(169) 00:13:27.060 fused_ordering(170) 00:13:27.060 fused_ordering(171) 00:13:27.060 fused_ordering(172) 00:13:27.060 fused_ordering(173) 00:13:27.060 fused_ordering(174) 00:13:27.060 fused_ordering(175) 00:13:27.060 fused_ordering(176) 00:13:27.060 fused_ordering(177) 00:13:27.060 fused_ordering(178) 00:13:27.060 fused_ordering(179) 00:13:27.060 fused_ordering(180) 00:13:27.060 fused_ordering(181) 00:13:27.060 fused_ordering(182) 00:13:27.060 fused_ordering(183) 00:13:27.060 fused_ordering(184) 00:13:27.060 fused_ordering(185) 00:13:27.060 fused_ordering(186) 00:13:27.060 fused_ordering(187) 00:13:27.060 fused_ordering(188) 00:13:27.060 fused_ordering(189) 00:13:27.060 fused_ordering(190) 00:13:27.060 fused_ordering(191) 00:13:27.060 fused_ordering(192) 00:13:27.060 fused_ordering(193) 00:13:27.060 fused_ordering(194) 00:13:27.060 fused_ordering(195) 00:13:27.060 fused_ordering(196) 00:13:27.060 fused_ordering(197) 00:13:27.060 fused_ordering(198) 00:13:27.060 fused_ordering(199) 00:13:27.060 fused_ordering(200) 00:13:27.060 fused_ordering(201) 00:13:27.060 fused_ordering(202) 00:13:27.060 fused_ordering(203) 00:13:27.060 fused_ordering(204) 00:13:27.060 fused_ordering(205) 00:13:27.322 fused_ordering(206) 00:13:27.322 fused_ordering(207) 00:13:27.322 fused_ordering(208) 00:13:27.322 fused_ordering(209) 00:13:27.322 fused_ordering(210) 00:13:27.322 fused_ordering(211) 00:13:27.322 fused_ordering(212) 00:13:27.322 fused_ordering(213) 00:13:27.322 fused_ordering(214) 00:13:27.322 fused_ordering(215) 00:13:27.322 fused_ordering(216) 00:13:27.322 fused_ordering(217) 00:13:27.322 fused_ordering(218) 00:13:27.322 fused_ordering(219) 00:13:27.322 fused_ordering(220) 00:13:27.322 fused_ordering(221) 00:13:27.322 fused_ordering(222) 00:13:27.322 fused_ordering(223) 00:13:27.322 fused_ordering(224) 00:13:27.322 fused_ordering(225) 00:13:27.322 fused_ordering(226) 00:13:27.322 fused_ordering(227) 00:13:27.322 fused_ordering(228) 00:13:27.322 fused_ordering(229) 00:13:27.322 fused_ordering(230) 00:13:27.322 fused_ordering(231) 00:13:27.322 fused_ordering(232) 00:13:27.322 fused_ordering(233) 00:13:27.322 fused_ordering(234) 00:13:27.322 fused_ordering(235) 00:13:27.322 fused_ordering(236) 00:13:27.322 fused_ordering(237) 00:13:27.322 fused_ordering(238) 00:13:27.322 fused_ordering(239) 00:13:27.322 fused_ordering(240) 00:13:27.322 fused_ordering(241) 00:13:27.322 fused_ordering(242) 00:13:27.322 fused_ordering(243) 00:13:27.322 fused_ordering(244) 00:13:27.322 fused_ordering(245) 00:13:27.322 fused_ordering(246) 00:13:27.322 fused_ordering(247) 00:13:27.322 fused_ordering(248) 00:13:27.322 fused_ordering(249) 00:13:27.322 fused_ordering(250) 00:13:27.322 fused_ordering(251) 00:13:27.322 fused_ordering(252) 00:13:27.322 fused_ordering(253) 00:13:27.322 fused_ordering(254) 00:13:27.322 fused_ordering(255) 00:13:27.322 fused_ordering(256) 00:13:27.322 fused_ordering(257) 00:13:27.322 fused_ordering(258) 00:13:27.322 fused_ordering(259) 00:13:27.322 fused_ordering(260) 00:13:27.322 fused_ordering(261) 00:13:27.322 fused_ordering(262) 00:13:27.322 fused_ordering(263) 00:13:27.322 fused_ordering(264) 00:13:27.322 fused_ordering(265) 00:13:27.322 fused_ordering(266) 00:13:27.322 fused_ordering(267) 00:13:27.322 fused_ordering(268) 00:13:27.322 fused_ordering(269) 00:13:27.322 fused_ordering(270) 00:13:27.322 fused_ordering(271) 00:13:27.322 fused_ordering(272) 00:13:27.322 fused_ordering(273) 00:13:27.322 fused_ordering(274) 00:13:27.322 fused_ordering(275) 00:13:27.322 fused_ordering(276) 00:13:27.322 fused_ordering(277) 00:13:27.322 fused_ordering(278) 00:13:27.322 fused_ordering(279) 00:13:27.322 fused_ordering(280) 00:13:27.322 fused_ordering(281) 00:13:27.322 fused_ordering(282) 00:13:27.322 fused_ordering(283) 00:13:27.322 fused_ordering(284) 00:13:27.322 fused_ordering(285) 00:13:27.322 fused_ordering(286) 00:13:27.322 fused_ordering(287) 00:13:27.322 fused_ordering(288) 00:13:27.323 fused_ordering(289) 00:13:27.323 fused_ordering(290) 00:13:27.323 fused_ordering(291) 00:13:27.323 fused_ordering(292) 00:13:27.323 fused_ordering(293) 00:13:27.323 fused_ordering(294) 00:13:27.323 fused_ordering(295) 00:13:27.323 fused_ordering(296) 00:13:27.323 fused_ordering(297) 00:13:27.323 fused_ordering(298) 00:13:27.323 fused_ordering(299) 00:13:27.323 fused_ordering(300) 00:13:27.323 fused_ordering(301) 00:13:27.323 fused_ordering(302) 00:13:27.323 fused_ordering(303) 00:13:27.323 fused_ordering(304) 00:13:27.323 fused_ordering(305) 00:13:27.323 fused_ordering(306) 00:13:27.323 fused_ordering(307) 00:13:27.323 fused_ordering(308) 00:13:27.323 fused_ordering(309) 00:13:27.323 fused_ordering(310) 00:13:27.323 fused_ordering(311) 00:13:27.323 fused_ordering(312) 00:13:27.323 fused_ordering(313) 00:13:27.323 fused_ordering(314) 00:13:27.323 fused_ordering(315) 00:13:27.323 fused_ordering(316) 00:13:27.323 fused_ordering(317) 00:13:27.323 fused_ordering(318) 00:13:27.323 fused_ordering(319) 00:13:27.323 fused_ordering(320) 00:13:27.323 fused_ordering(321) 00:13:27.323 fused_ordering(322) 00:13:27.323 fused_ordering(323) 00:13:27.323 fused_ordering(324) 00:13:27.323 fused_ordering(325) 00:13:27.323 fused_ordering(326) 00:13:27.323 fused_ordering(327) 00:13:27.323 fused_ordering(328) 00:13:27.323 fused_ordering(329) 00:13:27.323 fused_ordering(330) 00:13:27.323 fused_ordering(331) 00:13:27.323 fused_ordering(332) 00:13:27.323 fused_ordering(333) 00:13:27.323 fused_ordering(334) 00:13:27.323 fused_ordering(335) 00:13:27.323 fused_ordering(336) 00:13:27.323 fused_ordering(337) 00:13:27.323 fused_ordering(338) 00:13:27.323 fused_ordering(339) 00:13:27.323 fused_ordering(340) 00:13:27.323 fused_ordering(341) 00:13:27.323 fused_ordering(342) 00:13:27.323 fused_ordering(343) 00:13:27.323 fused_ordering(344) 00:13:27.323 fused_ordering(345) 00:13:27.323 fused_ordering(346) 00:13:27.323 fused_ordering(347) 00:13:27.323 fused_ordering(348) 00:13:27.323 fused_ordering(349) 00:13:27.323 fused_ordering(350) 00:13:27.323 fused_ordering(351) 00:13:27.323 fused_ordering(352) 00:13:27.323 fused_ordering(353) 00:13:27.323 fused_ordering(354) 00:13:27.323 fused_ordering(355) 00:13:27.323 fused_ordering(356) 00:13:27.323 fused_ordering(357) 00:13:27.323 fused_ordering(358) 00:13:27.323 fused_ordering(359) 00:13:27.323 fused_ordering(360) 00:13:27.323 fused_ordering(361) 00:13:27.323 fused_ordering(362) 00:13:27.323 fused_ordering(363) 00:13:27.323 fused_ordering(364) 00:13:27.323 fused_ordering(365) 00:13:27.323 fused_ordering(366) 00:13:27.323 fused_ordering(367) 00:13:27.323 fused_ordering(368) 00:13:27.323 fused_ordering(369) 00:13:27.323 fused_ordering(370) 00:13:27.323 fused_ordering(371) 00:13:27.323 fused_ordering(372) 00:13:27.323 fused_ordering(373) 00:13:27.323 fused_ordering(374) 00:13:27.323 fused_ordering(375) 00:13:27.323 fused_ordering(376) 00:13:27.323 fused_ordering(377) 00:13:27.323 fused_ordering(378) 00:13:27.323 fused_ordering(379) 00:13:27.323 fused_ordering(380) 00:13:27.323 fused_ordering(381) 00:13:27.323 fused_ordering(382) 00:13:27.323 fused_ordering(383) 00:13:27.323 fused_ordering(384) 00:13:27.323 fused_ordering(385) 00:13:27.323 fused_ordering(386) 00:13:27.323 fused_ordering(387) 00:13:27.323 fused_ordering(388) 00:13:27.323 fused_ordering(389) 00:13:27.323 fused_ordering(390) 00:13:27.323 fused_ordering(391) 00:13:27.323 fused_ordering(392) 00:13:27.323 fused_ordering(393) 00:13:27.323 fused_ordering(394) 00:13:27.323 fused_ordering(395) 00:13:27.323 fused_ordering(396) 00:13:27.323 fused_ordering(397) 00:13:27.323 fused_ordering(398) 00:13:27.323 fused_ordering(399) 00:13:27.323 fused_ordering(400) 00:13:27.323 fused_ordering(401) 00:13:27.323 fused_ordering(402) 00:13:27.323 fused_ordering(403) 00:13:27.323 fused_ordering(404) 00:13:27.323 fused_ordering(405) 00:13:27.323 fused_ordering(406) 00:13:27.323 fused_ordering(407) 00:13:27.323 fused_ordering(408) 00:13:27.323 fused_ordering(409) 00:13:27.323 fused_ordering(410) 00:13:27.895 fused_ordering(411) 00:13:27.895 fused_ordering(412) 00:13:27.895 fused_ordering(413) 00:13:27.895 fused_ordering(414) 00:13:27.895 fused_ordering(415) 00:13:27.895 fused_ordering(416) 00:13:27.895 fused_ordering(417) 00:13:27.895 fused_ordering(418) 00:13:27.895 fused_ordering(419) 00:13:27.895 fused_ordering(420) 00:13:27.895 fused_ordering(421) 00:13:27.895 fused_ordering(422) 00:13:27.895 fused_ordering(423) 00:13:27.895 fused_ordering(424) 00:13:27.895 fused_ordering(425) 00:13:27.895 fused_ordering(426) 00:13:27.895 fused_ordering(427) 00:13:27.895 fused_ordering(428) 00:13:27.895 fused_ordering(429) 00:13:27.895 fused_ordering(430) 00:13:27.895 fused_ordering(431) 00:13:27.896 fused_ordering(432) 00:13:27.896 fused_ordering(433) 00:13:27.896 fused_ordering(434) 00:13:27.896 fused_ordering(435) 00:13:27.896 fused_ordering(436) 00:13:27.896 fused_ordering(437) 00:13:27.896 fused_ordering(438) 00:13:27.896 fused_ordering(439) 00:13:27.896 fused_ordering(440) 00:13:27.896 fused_ordering(441) 00:13:27.896 fused_ordering(442) 00:13:27.896 fused_ordering(443) 00:13:27.896 fused_ordering(444) 00:13:27.896 fused_ordering(445) 00:13:27.896 fused_ordering(446) 00:13:27.896 fused_ordering(447) 00:13:27.896 fused_ordering(448) 00:13:27.896 fused_ordering(449) 00:13:27.896 fused_ordering(450) 00:13:27.896 fused_ordering(451) 00:13:27.896 fused_ordering(452) 00:13:27.896 fused_ordering(453) 00:13:27.896 fused_ordering(454) 00:13:27.896 fused_ordering(455) 00:13:27.896 fused_ordering(456) 00:13:27.896 fused_ordering(457) 00:13:27.896 fused_ordering(458) 00:13:27.896 fused_ordering(459) 00:13:27.896 fused_ordering(460) 00:13:27.896 fused_ordering(461) 00:13:27.896 fused_ordering(462) 00:13:27.896 fused_ordering(463) 00:13:27.896 fused_ordering(464) 00:13:27.896 fused_ordering(465) 00:13:27.896 fused_ordering(466) 00:13:27.896 fused_ordering(467) 00:13:27.896 fused_ordering(468) 00:13:27.896 fused_ordering(469) 00:13:27.896 fused_ordering(470) 00:13:27.896 fused_ordering(471) 00:13:27.896 fused_ordering(472) 00:13:27.896 fused_ordering(473) 00:13:27.896 fused_ordering(474) 00:13:27.896 fused_ordering(475) 00:13:27.896 fused_ordering(476) 00:13:27.896 fused_ordering(477) 00:13:27.896 fused_ordering(478) 00:13:27.896 fused_ordering(479) 00:13:27.896 fused_ordering(480) 00:13:27.896 fused_ordering(481) 00:13:27.896 fused_ordering(482) 00:13:27.896 fused_ordering(483) 00:13:27.896 fused_ordering(484) 00:13:27.896 fused_ordering(485) 00:13:27.896 fused_ordering(486) 00:13:27.896 fused_ordering(487) 00:13:27.896 fused_ordering(488) 00:13:27.896 fused_ordering(489) 00:13:27.896 fused_ordering(490) 00:13:27.896 fused_ordering(491) 00:13:27.896 fused_ordering(492) 00:13:27.896 fused_ordering(493) 00:13:27.896 fused_ordering(494) 00:13:27.896 fused_ordering(495) 00:13:27.896 fused_ordering(496) 00:13:27.896 fused_ordering(497) 00:13:27.896 fused_ordering(498) 00:13:27.896 fused_ordering(499) 00:13:27.896 fused_ordering(500) 00:13:27.896 fused_ordering(501) 00:13:27.896 fused_ordering(502) 00:13:27.896 fused_ordering(503) 00:13:27.896 fused_ordering(504) 00:13:27.896 fused_ordering(505) 00:13:27.896 fused_ordering(506) 00:13:27.896 fused_ordering(507) 00:13:27.896 fused_ordering(508) 00:13:27.896 fused_ordering(509) 00:13:27.896 fused_ordering(510) 00:13:27.896 fused_ordering(511) 00:13:27.896 fused_ordering(512) 00:13:27.896 fused_ordering(513) 00:13:27.896 fused_ordering(514) 00:13:27.896 fused_ordering(515) 00:13:27.896 fused_ordering(516) 00:13:27.896 fused_ordering(517) 00:13:27.896 fused_ordering(518) 00:13:27.896 fused_ordering(519) 00:13:27.896 fused_ordering(520) 00:13:27.896 fused_ordering(521) 00:13:27.896 fused_ordering(522) 00:13:27.896 fused_ordering(523) 00:13:27.896 fused_ordering(524) 00:13:27.896 fused_ordering(525) 00:13:27.896 fused_ordering(526) 00:13:27.896 fused_ordering(527) 00:13:27.896 fused_ordering(528) 00:13:27.896 fused_ordering(529) 00:13:27.896 fused_ordering(530) 00:13:27.896 fused_ordering(531) 00:13:27.896 fused_ordering(532) 00:13:27.896 fused_ordering(533) 00:13:27.896 fused_ordering(534) 00:13:27.896 fused_ordering(535) 00:13:27.896 fused_ordering(536) 00:13:27.896 fused_ordering(537) 00:13:27.896 fused_ordering(538) 00:13:27.896 fused_ordering(539) 00:13:27.896 fused_ordering(540) 00:13:27.896 fused_ordering(541) 00:13:27.896 fused_ordering(542) 00:13:27.896 fused_ordering(543) 00:13:27.896 fused_ordering(544) 00:13:27.896 fused_ordering(545) 00:13:27.896 fused_ordering(546) 00:13:27.896 fused_ordering(547) 00:13:27.896 fused_ordering(548) 00:13:27.896 fused_ordering(549) 00:13:27.896 fused_ordering(550) 00:13:27.896 fused_ordering(551) 00:13:27.896 fused_ordering(552) 00:13:27.896 fused_ordering(553) 00:13:27.896 fused_ordering(554) 00:13:27.896 fused_ordering(555) 00:13:27.896 fused_ordering(556) 00:13:27.896 fused_ordering(557) 00:13:27.896 fused_ordering(558) 00:13:27.896 fused_ordering(559) 00:13:27.896 fused_ordering(560) 00:13:27.896 fused_ordering(561) 00:13:27.896 fused_ordering(562) 00:13:27.896 fused_ordering(563) 00:13:27.896 fused_ordering(564) 00:13:27.896 fused_ordering(565) 00:13:27.896 fused_ordering(566) 00:13:27.896 fused_ordering(567) 00:13:27.896 fused_ordering(568) 00:13:27.896 fused_ordering(569) 00:13:27.896 fused_ordering(570) 00:13:27.896 fused_ordering(571) 00:13:27.896 fused_ordering(572) 00:13:27.896 fused_ordering(573) 00:13:27.896 fused_ordering(574) 00:13:27.896 fused_ordering(575) 00:13:27.896 fused_ordering(576) 00:13:27.896 fused_ordering(577) 00:13:27.896 fused_ordering(578) 00:13:27.896 fused_ordering(579) 00:13:27.896 fused_ordering(580) 00:13:27.896 fused_ordering(581) 00:13:27.896 fused_ordering(582) 00:13:27.896 fused_ordering(583) 00:13:27.896 fused_ordering(584) 00:13:27.896 fused_ordering(585) 00:13:27.896 fused_ordering(586) 00:13:27.896 fused_ordering(587) 00:13:27.896 fused_ordering(588) 00:13:27.896 fused_ordering(589) 00:13:27.896 fused_ordering(590) 00:13:27.896 fused_ordering(591) 00:13:27.896 fused_ordering(592) 00:13:27.896 fused_ordering(593) 00:13:27.896 fused_ordering(594) 00:13:27.896 fused_ordering(595) 00:13:27.896 fused_ordering(596) 00:13:27.896 fused_ordering(597) 00:13:27.896 fused_ordering(598) 00:13:27.896 fused_ordering(599) 00:13:27.896 fused_ordering(600) 00:13:27.896 fused_ordering(601) 00:13:27.896 fused_ordering(602) 00:13:27.896 fused_ordering(603) 00:13:27.896 fused_ordering(604) 00:13:27.896 fused_ordering(605) 00:13:27.896 fused_ordering(606) 00:13:27.896 fused_ordering(607) 00:13:27.896 fused_ordering(608) 00:13:27.896 fused_ordering(609) 00:13:27.896 fused_ordering(610) 00:13:27.896 fused_ordering(611) 00:13:27.896 fused_ordering(612) 00:13:27.896 fused_ordering(613) 00:13:27.896 fused_ordering(614) 00:13:27.896 fused_ordering(615) 00:13:28.156 fused_ordering(616) 00:13:28.156 fused_ordering(617) 00:13:28.156 fused_ordering(618) 00:13:28.156 fused_ordering(619) 00:13:28.156 fused_ordering(620) 00:13:28.156 fused_ordering(621) 00:13:28.156 fused_ordering(622) 00:13:28.156 fused_ordering(623) 00:13:28.156 fused_ordering(624) 00:13:28.156 fused_ordering(625) 00:13:28.156 fused_ordering(626) 00:13:28.156 fused_ordering(627) 00:13:28.156 fused_ordering(628) 00:13:28.156 fused_ordering(629) 00:13:28.156 fused_ordering(630) 00:13:28.156 fused_ordering(631) 00:13:28.156 fused_ordering(632) 00:13:28.156 fused_ordering(633) 00:13:28.156 fused_ordering(634) 00:13:28.156 fused_ordering(635) 00:13:28.156 fused_ordering(636) 00:13:28.156 fused_ordering(637) 00:13:28.156 fused_ordering(638) 00:13:28.156 fused_ordering(639) 00:13:28.156 fused_ordering(640) 00:13:28.156 fused_ordering(641) 00:13:28.156 fused_ordering(642) 00:13:28.156 fused_ordering(643) 00:13:28.156 fused_ordering(644) 00:13:28.156 fused_ordering(645) 00:13:28.156 fused_ordering(646) 00:13:28.156 fused_ordering(647) 00:13:28.156 fused_ordering(648) 00:13:28.156 fused_ordering(649) 00:13:28.156 fused_ordering(650) 00:13:28.156 fused_ordering(651) 00:13:28.156 fused_ordering(652) 00:13:28.156 fused_ordering(653) 00:13:28.156 fused_ordering(654) 00:13:28.156 fused_ordering(655) 00:13:28.156 fused_ordering(656) 00:13:28.156 fused_ordering(657) 00:13:28.156 fused_ordering(658) 00:13:28.156 fused_ordering(659) 00:13:28.156 fused_ordering(660) 00:13:28.156 fused_ordering(661) 00:13:28.156 fused_ordering(662) 00:13:28.156 fused_ordering(663) 00:13:28.156 fused_ordering(664) 00:13:28.156 fused_ordering(665) 00:13:28.156 fused_ordering(666) 00:13:28.156 fused_ordering(667) 00:13:28.156 fused_ordering(668) 00:13:28.156 fused_ordering(669) 00:13:28.157 fused_ordering(670) 00:13:28.157 fused_ordering(671) 00:13:28.157 fused_ordering(672) 00:13:28.157 fused_ordering(673) 00:13:28.157 fused_ordering(674) 00:13:28.157 fused_ordering(675) 00:13:28.157 fused_ordering(676) 00:13:28.157 fused_ordering(677) 00:13:28.157 fused_ordering(678) 00:13:28.157 fused_ordering(679) 00:13:28.157 fused_ordering(680) 00:13:28.157 fused_ordering(681) 00:13:28.157 fused_ordering(682) 00:13:28.157 fused_ordering(683) 00:13:28.157 fused_ordering(684) 00:13:28.157 fused_ordering(685) 00:13:28.157 fused_ordering(686) 00:13:28.157 fused_ordering(687) 00:13:28.157 fused_ordering(688) 00:13:28.157 fused_ordering(689) 00:13:28.157 fused_ordering(690) 00:13:28.157 fused_ordering(691) 00:13:28.157 fused_ordering(692) 00:13:28.157 fused_ordering(693) 00:13:28.157 fused_ordering(694) 00:13:28.157 fused_ordering(695) 00:13:28.157 fused_ordering(696) 00:13:28.157 fused_ordering(697) 00:13:28.157 fused_ordering(698) 00:13:28.157 fused_ordering(699) 00:13:28.157 fused_ordering(700) 00:13:28.157 fused_ordering(701) 00:13:28.157 fused_ordering(702) 00:13:28.157 fused_ordering(703) 00:13:28.157 fused_ordering(704) 00:13:28.157 fused_ordering(705) 00:13:28.157 fused_ordering(706) 00:13:28.157 fused_ordering(707) 00:13:28.157 fused_ordering(708) 00:13:28.157 fused_ordering(709) 00:13:28.157 fused_ordering(710) 00:13:28.157 fused_ordering(711) 00:13:28.157 fused_ordering(712) 00:13:28.157 fused_ordering(713) 00:13:28.157 fused_ordering(714) 00:13:28.157 fused_ordering(715) 00:13:28.157 fused_ordering(716) 00:13:28.157 fused_ordering(717) 00:13:28.157 fused_ordering(718) 00:13:28.157 fused_ordering(719) 00:13:28.157 fused_ordering(720) 00:13:28.157 fused_ordering(721) 00:13:28.157 fused_ordering(722) 00:13:28.157 fused_ordering(723) 00:13:28.157 fused_ordering(724) 00:13:28.157 fused_ordering(725) 00:13:28.157 fused_ordering(726) 00:13:28.157 fused_ordering(727) 00:13:28.157 fused_ordering(728) 00:13:28.157 fused_ordering(729) 00:13:28.157 fused_ordering(730) 00:13:28.157 fused_ordering(731) 00:13:28.157 fused_ordering(732) 00:13:28.157 fused_ordering(733) 00:13:28.157 fused_ordering(734) 00:13:28.157 fused_ordering(735) 00:13:28.157 fused_ordering(736) 00:13:28.157 fused_ordering(737) 00:13:28.157 fused_ordering(738) 00:13:28.157 fused_ordering(739) 00:13:28.157 fused_ordering(740) 00:13:28.157 fused_ordering(741) 00:13:28.157 fused_ordering(742) 00:13:28.157 fused_ordering(743) 00:13:28.157 fused_ordering(744) 00:13:28.157 fused_ordering(745) 00:13:28.157 fused_ordering(746) 00:13:28.157 fused_ordering(747) 00:13:28.157 fused_ordering(748) 00:13:28.157 fused_ordering(749) 00:13:28.157 fused_ordering(750) 00:13:28.157 fused_ordering(751) 00:13:28.157 fused_ordering(752) 00:13:28.157 fused_ordering(753) 00:13:28.157 fused_ordering(754) 00:13:28.157 fused_ordering(755) 00:13:28.157 fused_ordering(756) 00:13:28.157 fused_ordering(757) 00:13:28.157 fused_ordering(758) 00:13:28.157 fused_ordering(759) 00:13:28.157 fused_ordering(760) 00:13:28.157 fused_ordering(761) 00:13:28.157 fused_ordering(762) 00:13:28.157 fused_ordering(763) 00:13:28.157 fused_ordering(764) 00:13:28.157 fused_ordering(765) 00:13:28.157 fused_ordering(766) 00:13:28.157 fused_ordering(767) 00:13:28.157 fused_ordering(768) 00:13:28.157 fused_ordering(769) 00:13:28.157 fused_ordering(770) 00:13:28.157 fused_ordering(771) 00:13:28.157 fused_ordering(772) 00:13:28.157 fused_ordering(773) 00:13:28.157 fused_ordering(774) 00:13:28.157 fused_ordering(775) 00:13:28.157 fused_ordering(776) 00:13:28.157 fused_ordering(777) 00:13:28.157 fused_ordering(778) 00:13:28.157 fused_ordering(779) 00:13:28.157 fused_ordering(780) 00:13:28.157 fused_ordering(781) 00:13:28.157 fused_ordering(782) 00:13:28.157 fused_ordering(783) 00:13:28.157 fused_ordering(784) 00:13:28.157 fused_ordering(785) 00:13:28.157 fused_ordering(786) 00:13:28.157 fused_ordering(787) 00:13:28.157 fused_ordering(788) 00:13:28.157 fused_ordering(789) 00:13:28.157 fused_ordering(790) 00:13:28.157 fused_ordering(791) 00:13:28.157 fused_ordering(792) 00:13:28.157 fused_ordering(793) 00:13:28.157 fused_ordering(794) 00:13:28.157 fused_ordering(795) 00:13:28.157 fused_ordering(796) 00:13:28.157 fused_ordering(797) 00:13:28.157 fused_ordering(798) 00:13:28.157 fused_ordering(799) 00:13:28.157 fused_ordering(800) 00:13:28.157 fused_ordering(801) 00:13:28.157 fused_ordering(802) 00:13:28.157 fused_ordering(803) 00:13:28.157 fused_ordering(804) 00:13:28.157 fused_ordering(805) 00:13:28.157 fused_ordering(806) 00:13:28.157 fused_ordering(807) 00:13:28.157 fused_ordering(808) 00:13:28.157 fused_ordering(809) 00:13:28.157 fused_ordering(810) 00:13:28.157 fused_ordering(811) 00:13:28.157 fused_ordering(812) 00:13:28.157 fused_ordering(813) 00:13:28.157 fused_ordering(814) 00:13:28.157 fused_ordering(815) 00:13:28.157 fused_ordering(816) 00:13:28.157 fused_ordering(817) 00:13:28.157 fused_ordering(818) 00:13:28.157 fused_ordering(819) 00:13:28.157 fused_ordering(820) 00:13:29.100 fused_o[2024-10-17 17:34:36.658821] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10934f0 is same with the state(6) to be set 00:13:29.100 rdering(821) 00:13:29.100 fused_ordering(822) 00:13:29.100 fused_ordering(823) 00:13:29.100 fused_ordering(824) 00:13:29.100 fused_ordering(825) 00:13:29.100 fused_ordering(826) 00:13:29.100 fused_ordering(827) 00:13:29.100 fused_ordering(828) 00:13:29.100 fused_ordering(829) 00:13:29.100 fused_ordering(830) 00:13:29.100 fused_ordering(831) 00:13:29.100 fused_ordering(832) 00:13:29.100 fused_ordering(833) 00:13:29.100 fused_ordering(834) 00:13:29.100 fused_ordering(835) 00:13:29.100 fused_ordering(836) 00:13:29.100 fused_ordering(837) 00:13:29.100 fused_ordering(838) 00:13:29.100 fused_ordering(839) 00:13:29.100 fused_ordering(840) 00:13:29.100 fused_ordering(841) 00:13:29.100 fused_ordering(842) 00:13:29.100 fused_ordering(843) 00:13:29.100 fused_ordering(844) 00:13:29.100 fused_ordering(845) 00:13:29.100 fused_ordering(846) 00:13:29.100 fused_ordering(847) 00:13:29.100 fused_ordering(848) 00:13:29.100 fused_ordering(849) 00:13:29.100 fused_ordering(850) 00:13:29.100 fused_ordering(851) 00:13:29.100 fused_ordering(852) 00:13:29.100 fused_ordering(853) 00:13:29.100 fused_ordering(854) 00:13:29.100 fused_ordering(855) 00:13:29.100 fused_ordering(856) 00:13:29.100 fused_ordering(857) 00:13:29.100 fused_ordering(858) 00:13:29.100 fused_ordering(859) 00:13:29.100 fused_ordering(860) 00:13:29.100 fused_ordering(861) 00:13:29.100 fused_ordering(862) 00:13:29.100 fused_ordering(863) 00:13:29.100 fused_ordering(864) 00:13:29.100 fused_ordering(865) 00:13:29.100 fused_ordering(866) 00:13:29.100 fused_ordering(867) 00:13:29.100 fused_ordering(868) 00:13:29.100 fused_ordering(869) 00:13:29.100 fused_ordering(870) 00:13:29.100 fused_ordering(871) 00:13:29.100 fused_ordering(872) 00:13:29.100 fused_ordering(873) 00:13:29.100 fused_ordering(874) 00:13:29.100 fused_ordering(875) 00:13:29.100 fused_ordering(876) 00:13:29.100 fused_ordering(877) 00:13:29.100 fused_ordering(878) 00:13:29.100 fused_ordering(879) 00:13:29.100 fused_ordering(880) 00:13:29.100 fused_ordering(881) 00:13:29.100 fused_ordering(882) 00:13:29.100 fused_ordering(883) 00:13:29.100 fused_ordering(884) 00:13:29.100 fused_ordering(885) 00:13:29.100 fused_ordering(886) 00:13:29.100 fused_ordering(887) 00:13:29.100 fused_ordering(888) 00:13:29.100 fused_ordering(889) 00:13:29.100 fused_ordering(890) 00:13:29.100 fused_ordering(891) 00:13:29.100 fused_ordering(892) 00:13:29.100 fused_ordering(893) 00:13:29.100 fused_ordering(894) 00:13:29.100 fused_ordering(895) 00:13:29.100 fused_ordering(896) 00:13:29.100 fused_ordering(897) 00:13:29.100 fused_ordering(898) 00:13:29.100 fused_ordering(899) 00:13:29.100 fused_ordering(900) 00:13:29.100 fused_ordering(901) 00:13:29.100 fused_ordering(902) 00:13:29.100 fused_ordering(903) 00:13:29.100 fused_ordering(904) 00:13:29.100 fused_ordering(905) 00:13:29.100 fused_ordering(906) 00:13:29.100 fused_ordering(907) 00:13:29.100 fused_ordering(908) 00:13:29.100 fused_ordering(909) 00:13:29.100 fused_ordering(910) 00:13:29.100 fused_ordering(911) 00:13:29.100 fused_ordering(912) 00:13:29.100 fused_ordering(913) 00:13:29.100 fused_ordering(914) 00:13:29.100 fused_ordering(915) 00:13:29.100 fused_ordering(916) 00:13:29.100 fused_ordering(917) 00:13:29.100 fused_ordering(918) 00:13:29.100 fused_ordering(919) 00:13:29.100 fused_ordering(920) 00:13:29.100 fused_ordering(921) 00:13:29.100 fused_ordering(922) 00:13:29.100 fused_ordering(923) 00:13:29.100 fused_ordering(924) 00:13:29.100 fused_ordering(925) 00:13:29.100 fused_ordering(926) 00:13:29.100 fused_ordering(927) 00:13:29.100 fused_ordering(928) 00:13:29.100 fused_ordering(929) 00:13:29.100 fused_ordering(930) 00:13:29.100 fused_ordering(931) 00:13:29.100 fused_ordering(932) 00:13:29.100 fused_ordering(933) 00:13:29.100 fused_ordering(934) 00:13:29.100 fused_ordering(935) 00:13:29.100 fused_ordering(936) 00:13:29.100 fused_ordering(937) 00:13:29.100 fused_ordering(938) 00:13:29.100 fused_ordering(939) 00:13:29.100 fused_ordering(940) 00:13:29.100 fused_ordering(941) 00:13:29.100 fused_ordering(942) 00:13:29.100 fused_ordering(943) 00:13:29.100 fused_ordering(944) 00:13:29.100 fused_ordering(945) 00:13:29.100 fused_ordering(946) 00:13:29.100 fused_ordering(947) 00:13:29.100 fused_ordering(948) 00:13:29.100 fused_ordering(949) 00:13:29.100 fused_ordering(950) 00:13:29.100 fused_ordering(951) 00:13:29.100 fused_ordering(952) 00:13:29.100 fused_ordering(953) 00:13:29.100 fused_ordering(954) 00:13:29.100 fused_ordering(955) 00:13:29.100 fused_ordering(956) 00:13:29.100 fused_ordering(957) 00:13:29.100 fused_ordering(958) 00:13:29.100 fused_ordering(959) 00:13:29.100 fused_ordering(960) 00:13:29.100 fused_ordering(961) 00:13:29.100 fused_ordering(962) 00:13:29.100 fused_ordering(963) 00:13:29.100 fused_ordering(964) 00:13:29.100 fused_ordering(965) 00:13:29.100 fused_ordering(966) 00:13:29.100 fused_ordering(967) 00:13:29.100 fused_ordering(968) 00:13:29.100 fused_ordering(969) 00:13:29.100 fused_ordering(970) 00:13:29.100 fused_ordering(971) 00:13:29.100 fused_ordering(972) 00:13:29.100 fused_ordering(973) 00:13:29.100 fused_ordering(974) 00:13:29.100 fused_ordering(975) 00:13:29.100 fused_ordering(976) 00:13:29.100 fused_ordering(977) 00:13:29.100 fused_ordering(978) 00:13:29.100 fused_ordering(979) 00:13:29.100 fused_ordering(980) 00:13:29.100 fused_ordering(981) 00:13:29.100 fused_ordering(982) 00:13:29.100 fused_ordering(983) 00:13:29.100 fused_ordering(984) 00:13:29.100 fused_ordering(985) 00:13:29.100 fused_ordering(986) 00:13:29.100 fused_ordering(987) 00:13:29.100 fused_ordering(988) 00:13:29.100 fused_ordering(989) 00:13:29.100 fused_ordering(990) 00:13:29.100 fused_ordering(991) 00:13:29.100 fused_ordering(992) 00:13:29.100 fused_ordering(993) 00:13:29.100 fused_ordering(994) 00:13:29.100 fused_ordering(995) 00:13:29.100 fused_ordering(996) 00:13:29.100 fused_ordering(997) 00:13:29.100 fused_ordering(998) 00:13:29.100 fused_ordering(999) 00:13:29.100 fused_ordering(1000) 00:13:29.100 fused_ordering(1001) 00:13:29.100 fused_ordering(1002) 00:13:29.100 fused_ordering(1003) 00:13:29.100 fused_ordering(1004) 00:13:29.100 fused_ordering(1005) 00:13:29.100 fused_ordering(1006) 00:13:29.100 fused_ordering(1007) 00:13:29.100 fused_ordering(1008) 00:13:29.100 fused_ordering(1009) 00:13:29.100 fused_ordering(1010) 00:13:29.100 fused_ordering(1011) 00:13:29.100 fused_ordering(1012) 00:13:29.100 fused_ordering(1013) 00:13:29.100 fused_ordering(1014) 00:13:29.100 fused_ordering(1015) 00:13:29.100 fused_ordering(1016) 00:13:29.100 fused_ordering(1017) 00:13:29.100 fused_ordering(1018) 00:13:29.100 fused_ordering(1019) 00:13:29.100 fused_ordering(1020) 00:13:29.100 fused_ordering(1021) 00:13:29.100 fused_ordering(1022) 00:13:29.100 fused_ordering(1023) 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@23 -- # trap - SIGINT SIGTERM EXIT 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@25 -- # nvmftestfini 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@514 -- # nvmfcleanup 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@121 -- # sync 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@124 -- # set +e 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:29.100 rmmod nvme_tcp 00:13:29.100 rmmod nvme_fabrics 00:13:29.100 rmmod nvme_keyring 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@128 -- # set -e 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@129 -- # return 0 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@515 -- # '[' -n 4176938 ']' 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@516 -- # killprocess 4176938 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@950 -- # '[' -z 4176938 ']' 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@954 -- # kill -0 4176938 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@955 -- # uname 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:29.100 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4176938 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4176938' 00:13:29.101 killing process with pid 4176938 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@969 -- # kill 4176938 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@974 -- # wait 4176938 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@297 -- # iptr 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@789 -- # iptables-save 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@789 -- # iptables-restore 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:29.101 17:34:36 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:31.649 00:13:31.649 real 0m13.554s 00:13:31.649 user 0m6.985s 00:13:31.649 sys 0m7.309s 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:13:31.649 ************************************ 00:13:31.649 END TEST nvmf_fused_ordering 00:13:31.649 ************************************ 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@26 -- # run_test nvmf_ns_masking test/nvmf/target/ns_masking.sh --transport=tcp 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:13:31.649 ************************************ 00:13:31.649 START TEST nvmf_ns_masking 00:13:31.649 ************************************ 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1125 -- # test/nvmf/target/ns_masking.sh --transport=tcp 00:13:31.649 * Looking for test storage... 00:13:31.649 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1691 -- # lcov --version 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # IFS=.-: 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # read -ra ver1 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # IFS=.-: 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # read -ra ver2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@338 -- # local 'op=<' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@340 -- # ver1_l=2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@341 -- # ver2_l=1 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@344 -- # case "$op" in 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@345 -- # : 1 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # decimal 1 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=1 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 1 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # ver1[v]=1 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # decimal 2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # ver2[v]=2 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # return 0 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:13:31.649 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:31.649 --rc genhtml_branch_coverage=1 00:13:31.649 --rc genhtml_function_coverage=1 00:13:31.649 --rc genhtml_legend=1 00:13:31.649 --rc geninfo_all_blocks=1 00:13:31.649 --rc geninfo_unexecuted_blocks=1 00:13:31.649 00:13:31.649 ' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:13:31.649 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:31.649 --rc genhtml_branch_coverage=1 00:13:31.649 --rc genhtml_function_coverage=1 00:13:31.649 --rc genhtml_legend=1 00:13:31.649 --rc geninfo_all_blocks=1 00:13:31.649 --rc geninfo_unexecuted_blocks=1 00:13:31.649 00:13:31.649 ' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:13:31.649 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:31.649 --rc genhtml_branch_coverage=1 00:13:31.649 --rc genhtml_function_coverage=1 00:13:31.649 --rc genhtml_legend=1 00:13:31.649 --rc geninfo_all_blocks=1 00:13:31.649 --rc geninfo_unexecuted_blocks=1 00:13:31.649 00:13:31.649 ' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:13:31.649 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:31.649 --rc genhtml_branch_coverage=1 00:13:31.649 --rc genhtml_function_coverage=1 00:13:31.649 --rc genhtml_legend=1 00:13:31.649 --rc geninfo_all_blocks=1 00:13:31.649 --rc geninfo_unexecuted_blocks=1 00:13:31.649 00:13:31.649 ' 00:13:31.649 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # uname -s 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@15 -- # shopt -s extglob 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@5 -- # export PATH 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@51 -- # : 0 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:31.650 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@10 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@11 -- # hostsock=/var/tmp/host.sock 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@12 -- # loops=5 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # uuidgen 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # ns1uuid=31ba59a7-462f-4ffc-a603-1967cd619aed 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # uuidgen 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # ns2uuid=eeb7382f-d91a-4445-87fd-cfeaa5156476 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@16 -- # SUBSYSNQN=nqn.2016-06.io.spdk:cnode1 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@17 -- # HOSTNQN1=nqn.2016-06.io.spdk:host1 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@18 -- # HOSTNQN2=nqn.2016-06.io.spdk:host2 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # uuidgen 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # HOSTID=9fe24eb4-9cc2-40a0-9e16-dc0d9da36cad 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@50 -- # nvmftestinit 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@474 -- # prepare_net_devs 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@436 -- # local -g is_hw=no 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@438 -- # remove_spdk_ns 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@309 -- # xtrace_disable 00:13:31.650 17:34:39 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # pci_devs=() 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # net_devs=() 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # e810=() 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # local -ga e810 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # x722=() 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # local -ga x722 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # mlx=() 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # local -ga mlx 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:13:39.799 Found 0000:31:00.0 (0x8086 - 0x159b) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:13:39.799 Found 0000:31:00.1 (0x8086 - 0x159b) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:39.799 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:13:39.799 Found net devices under 0000:31:00.0: cvl_0_0 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:13:39.800 Found net devices under 0000:31:00.1: cvl_0_1 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # is_hw=yes 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:39.800 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:39.800 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.640 ms 00:13:39.800 00:13:39.800 --- 10.0.0.2 ping statistics --- 00:13:39.800 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:39.800 rtt min/avg/max/mdev = 0.640/0.640/0.640/0.000 ms 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:39.800 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:39.800 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.280 ms 00:13:39.800 00:13:39.800 --- 10.0.0.1 ping statistics --- 00:13:39.800 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:39.800 rtt min/avg/max/mdev = 0.280/0.280/0.280/0.000 ms 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@448 -- # return 0 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:13:39.800 17:34:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@51 -- # nvmfappstart 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@724 -- # xtrace_disable 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@507 -- # nvmfpid=4181718 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@508 -- # waitforlisten 4181718 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@831 -- # '[' -z 4181718 ']' 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:39.800 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:13:39.800 [2024-10-17 17:34:47.080625] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:13:39.800 [2024-10-17 17:34:47.080702] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:39.800 [2024-10-17 17:34:47.146447] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:39.800 [2024-10-17 17:34:47.192673] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:39.800 [2024-10-17 17:34:47.192733] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:39.800 [2024-10-17 17:34:47.192741] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:39.800 [2024-10-17 17:34:47.192751] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:39.800 [2024-10-17 17:34:47.192756] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:39.800 [2024-10-17 17:34:47.193484] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # return 0 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@730 -- # xtrace_disable 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:13:39.800 [2024-10-17 17:34:47.505890] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@55 -- # MALLOC_BDEV_SIZE=64 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@56 -- # MALLOC_BLOCK_SIZE=512 00:13:39.800 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:13:39.800 Malloc1 00:13:40.062 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:13:40.062 Malloc2 00:13:40.062 17:34:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:13:40.323 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 00:13:40.583 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:40.583 [2024-10-17 17:34:48.485803] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:40.844 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@67 -- # connect 00:13:40.844 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 9fe24eb4-9cc2-40a0-9e16-dc0d9da36cad -a 10.0.0.2 -s 4420 -i 4 00:13:40.844 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 00:13:40.844 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:13:40.844 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:13:40.844 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:13:40.844 17:34:48 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:13:42.758 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:13:42.758 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:13:42.758 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@68 -- # ns_is_visible 0x1 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:43.018 [ 0]:0x1 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=7d2cfdbf69134a34a1d7a615f987431e 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 7d2cfdbf69134a34a1d7a615f987431e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:43.018 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@72 -- # ns_is_visible 0x1 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:43.279 [ 0]:0x1 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=7d2cfdbf69134a34a1d7a615f987431e 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 7d2cfdbf69134a34a1d7a615f987431e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@73 -- # ns_is_visible 0x2 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:43.279 17:34:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:13:43.279 [ 1]:0x2 00:13:43.279 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:43.279 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:43.279 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ee39fcb1140d4dc1b49108db60218017 00:13:43.279 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ee39fcb1140d4dc1b49108db60218017 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:43.279 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@75 -- # disconnect 00:13:43.279 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:43.539 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:43.539 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:43.800 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 --no-auto-visible 00:13:43.800 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@83 -- # connect 1 00:13:43.800 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 9fe24eb4-9cc2-40a0-9e16-dc0d9da36cad -a 10.0.0.2 -s 4420 -i 4 00:13:44.059 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 1 00:13:44.059 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:13:44.059 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:13:44.059 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n 1 ]] 00:13:44.059 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1201 -- # nvme_device_counter=1 00:13:44.059 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:13:46.600 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@84 -- # NOT ns_is_visible 0x1 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:46.601 17:34:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@85 -- # ns_is_visible 0x2 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:46.601 [ 0]:0x2 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ee39fcb1140d4dc1b49108db60218017 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ee39fcb1140d4dc1b49108db60218017 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@89 -- # ns_is_visible 0x1 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:46.601 [ 0]:0x1 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=7d2cfdbf69134a34a1d7a615f987431e 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 7d2cfdbf69134a34a1d7a615f987431e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@90 -- # ns_is_visible 0x2 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:13:46.601 [ 1]:0x2 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ee39fcb1140d4dc1b49108db60218017 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ee39fcb1140d4dc1b49108db60218017 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:46.601 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@94 -- # NOT ns_is_visible 0x1 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@95 -- # ns_is_visible 0x2 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:13:46.861 [ 0]:0x2 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ee39fcb1140d4dc1b49108db60218017 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ee39fcb1140d4dc1b49108db60218017 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@97 -- # disconnect 00:13:46.861 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:47.121 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:47.121 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:47.121 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@101 -- # connect 2 00:13:47.121 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 9fe24eb4-9cc2-40a0-9e16-dc0d9da36cad -a 10.0.0.2 -s 4420 -i 4 00:13:47.380 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 2 00:13:47.380 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:13:47.380 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:13:47.380 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n 2 ]] 00:13:47.381 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1201 -- # nvme_device_counter=2 00:13:47.381 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=2 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:13:49.290 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@102 -- # ns_is_visible 0x1 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:49.550 [ 0]:0x1 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=7d2cfdbf69134a34a1d7a615f987431e 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 7d2cfdbf69134a34a1d7a615f987431e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@103 -- # ns_is_visible 0x2 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:13:49.550 [ 1]:0x2 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ee39fcb1140d4dc1b49108db60218017 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ee39fcb1140d4dc1b49108db60218017 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:49.550 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@106 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@107 -- # NOT ns_is_visible 0x1 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:49.810 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@108 -- # ns_is_visible 0x2 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:13:49.811 [ 0]:0x2 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ee39fcb1140d4dc1b49108db60218017 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ee39fcb1140d4dc1b49108db60218017 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@111 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:13:49.811 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:13:50.072 [2024-10-17 17:34:57.831768] nvmf_rpc.c:1870:nvmf_rpc_ns_visible_paused: *ERROR*: Unable to add/remove nqn.2016-06.io.spdk:host1 to namespace ID 2 00:13:50.072 request: 00:13:50.072 { 00:13:50.072 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:13:50.072 "nsid": 2, 00:13:50.072 "host": "nqn.2016-06.io.spdk:host1", 00:13:50.072 "method": "nvmf_ns_remove_host", 00:13:50.072 "req_id": 1 00:13:50.072 } 00:13:50.072 Got JSON-RPC error response 00:13:50.072 response: 00:13:50.072 { 00:13:50.072 "code": -32602, 00:13:50.072 "message": "Invalid parameters" 00:13:50.072 } 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@112 -- # NOT ns_is_visible 0x1 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@113 -- # ns_is_visible 0x2 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:13:50.072 [ 0]:0x2 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=ee39fcb1140d4dc1b49108db60218017 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ ee39fcb1140d4dc1b49108db60218017 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@114 -- # disconnect 00:13:50.072 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:50.333 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@118 -- # hostpid=4184085 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@119 -- # trap 'killprocess $hostpid; nvmftestfini' SIGINT SIGTERM EXIT 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@117 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -r /var/tmp/host.sock -m 2 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@121 -- # waitforlisten 4184085 /var/tmp/host.sock 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@831 -- # '[' -z 4184085 ']' 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/host.sock 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:13:50.333 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:50.333 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:13:50.333 [2024-10-17 17:34:58.085247] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:13:50.333 [2024-10-17 17:34:58.085296] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4184085 ] 00:13:50.333 [2024-10-17 17:34:58.162593] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:50.333 [2024-10-17 17:34:58.198814] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:51.274 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:51.274 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # return 0 00:13:51.274 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@122 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:51.274 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:13:51.535 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # uuid2nguid 31ba59a7-462f-4ffc-a603-1967cd619aed 00:13:51.535 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@785 -- # tr -d - 00:13:51.535 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g 31BA59A7462F4FFCA6031967CD619AED -i 00:13:51.535 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # uuid2nguid eeb7382f-d91a-4445-87fd-cfeaa5156476 00:13:51.535 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@785 -- # tr -d - 00:13:51.535 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 -g EEB7382FD91A444587FDCFEAA5156476 -i 00:13:51.795 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:52.056 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@127 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host2 00:13:52.316 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@129 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:13:52.316 17:34:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:13:52.576 nvme0n1 00:13:52.576 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@131 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:13:52.576 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:13:52.836 nvme1n2 00:13:52.836 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # hostrpc bdev_get_bdevs 00:13:52.836 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # jq -r '.[].name' 00:13:52.836 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:13:52.836 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # sort 00:13:52.836 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # xargs 00:13:53.096 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # [[ nvme0n1 nvme1n2 == \n\v\m\e\0\n\1\ \n\v\m\e\1\n\2 ]] 00:13:53.096 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # hostrpc bdev_get_bdevs -b nvme0n1 00:13:53.096 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # jq -r '.[].uuid' 00:13:53.096 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme0n1 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # [[ 31ba59a7-462f-4ffc-a603-1967cd619aed == \3\1\b\a\5\9\a\7\-\4\6\2\f\-\4\f\f\c\-\a\6\0\3\-\1\9\6\7\c\d\6\1\9\a\e\d ]] 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # hostrpc bdev_get_bdevs -b nvme1n2 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # jq -r '.[].uuid' 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme1n2 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # [[ eeb7382f-d91a-4445-87fd-cfeaa5156476 == \e\e\b\7\3\8\2\f\-\d\9\1\a\-\4\4\4\5\-\8\7\f\d\-\c\f\e\a\a\5\1\5\6\4\7\6 ]] 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@138 -- # killprocess 4184085 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@950 -- # '[' -z 4184085 ']' 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # kill -0 4184085 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # uname 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:53.357 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4184085 00:13:53.619 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:13:53.619 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:13:53.619 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4184085' 00:13:53.619 killing process with pid 4184085 00:13:53.619 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@969 -- # kill 4184085 00:13:53.619 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@974 -- # wait 4184085 00:13:53.619 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@139 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@141 -- # trap - SIGINT SIGTERM EXIT 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@142 -- # nvmftestfini 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@514 -- # nvmfcleanup 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@121 -- # sync 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@124 -- # set +e 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:53.879 rmmod nvme_tcp 00:13:53.879 rmmod nvme_fabrics 00:13:53.879 rmmod nvme_keyring 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@128 -- # set -e 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@129 -- # return 0 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@515 -- # '[' -n 4181718 ']' 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@516 -- # killprocess 4181718 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@950 -- # '[' -z 4181718 ']' 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # kill -0 4181718 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # uname 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:53.879 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4181718 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4181718' 00:13:54.140 killing process with pid 4181718 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@969 -- # kill 4181718 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@974 -- # wait 4181718 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@297 -- # iptr 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@789 -- # iptables-save 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@789 -- # iptables-restore 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:54.140 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:56.683 00:13:56.683 real 0m24.875s 00:13:56.683 user 0m25.160s 00:13:56.683 sys 0m7.988s 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:13:56.683 ************************************ 00:13:56.683 END TEST nvmf_ns_masking 00:13:56.683 ************************************ 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@27 -- # [[ 1 -eq 1 ]] 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@28 -- # run_test nvmf_nvme_cli /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:13:56.683 ************************************ 00:13:56.683 START TEST nvmf_nvme_cli 00:13:56.683 ************************************ 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:13:56.683 * Looking for test storage... 00:13:56.683 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1691 -- # lcov --version 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # IFS=.-: 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # read -ra ver1 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # IFS=.-: 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # read -ra ver2 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@338 -- # local 'op=<' 00:13:56.683 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@340 -- # ver1_l=2 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@341 -- # ver2_l=1 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@344 -- # case "$op" in 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@345 -- # : 1 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # decimal 1 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=1 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 1 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # ver1[v]=1 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # decimal 2 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=2 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 2 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # ver2[v]=2 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # return 0 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:13:56.684 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:56.684 --rc genhtml_branch_coverage=1 00:13:56.684 --rc genhtml_function_coverage=1 00:13:56.684 --rc genhtml_legend=1 00:13:56.684 --rc geninfo_all_blocks=1 00:13:56.684 --rc geninfo_unexecuted_blocks=1 00:13:56.684 00:13:56.684 ' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:13:56.684 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:56.684 --rc genhtml_branch_coverage=1 00:13:56.684 --rc genhtml_function_coverage=1 00:13:56.684 --rc genhtml_legend=1 00:13:56.684 --rc geninfo_all_blocks=1 00:13:56.684 --rc geninfo_unexecuted_blocks=1 00:13:56.684 00:13:56.684 ' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:13:56.684 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:56.684 --rc genhtml_branch_coverage=1 00:13:56.684 --rc genhtml_function_coverage=1 00:13:56.684 --rc genhtml_legend=1 00:13:56.684 --rc geninfo_all_blocks=1 00:13:56.684 --rc geninfo_unexecuted_blocks=1 00:13:56.684 00:13:56.684 ' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:13:56.684 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:56.684 --rc genhtml_branch_coverage=1 00:13:56.684 --rc genhtml_function_coverage=1 00:13:56.684 --rc genhtml_legend=1 00:13:56.684 --rc geninfo_all_blocks=1 00:13:56.684 --rc geninfo_unexecuted_blocks=1 00:13:56.684 00:13:56.684 ' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # uname -s 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@15 -- # shopt -s extglob 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@5 -- # export PATH 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@51 -- # : 0 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:56.684 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@14 -- # devs=() 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@16 -- # nvmftestinit 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@474 -- # prepare_net_devs 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@436 -- # local -g is_hw=no 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@438 -- # remove_spdk_ns 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:13:56.684 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:13:56.685 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@309 -- # xtrace_disable 00:13:56.685 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # pci_devs=() 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # net_devs=() 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # e810=() 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # local -ga e810 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # x722=() 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # local -ga x722 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # mlx=() 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # local -ga mlx 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:14:04.845 Found 0000:31:00.0 (0x8086 - 0x159b) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:14:04.845 Found 0000:31:00.1 (0x8086 - 0x159b) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ up == up ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:14:04.845 Found net devices under 0000:31:00.0: cvl_0_0 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ up == up ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:14:04.845 Found net devices under 0000:31:00.1: cvl_0_1 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # is_hw=yes 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:14:04.845 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:04.846 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:04.846 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.392 ms 00:14:04.846 00:14:04.846 --- 10.0.0.2 ping statistics --- 00:14:04.846 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:04.846 rtt min/avg/max/mdev = 0.392/0.392/0.392/0.000 ms 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:04.846 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:04.846 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.293 ms 00:14:04.846 00:14:04.846 --- 10.0.0.1 ping statistics --- 00:14:04.846 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:04.846 rtt min/avg/max/mdev = 0.293/0.293/0.293/0.000 ms 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@448 -- # return 0 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@17 -- # nvmfappstart -m 0xF 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@724 -- # xtrace_disable 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@507 -- # nvmfpid=4189069 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@508 -- # waitforlisten 4189069 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@831 -- # '[' -z 4189069 ']' 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:04.846 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:04.846 17:35:11 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:04.846 [2024-10-17 17:35:11.972371] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:14:04.846 [2024-10-17 17:35:11.972436] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:04.846 [2024-10-17 17:35:12.062520] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:04.846 [2024-10-17 17:35:12.117583] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:04.846 [2024-10-17 17:35:12.117636] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:04.846 [2024-10-17 17:35:12.117644] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:04.846 [2024-10-17 17:35:12.117651] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:04.846 [2024-10-17 17:35:12.117657] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:04.846 [2024-10-17 17:35:12.119805] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:04.846 [2024-10-17 17:35:12.120019] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:04.846 [2024-10-17 17:35:12.120019] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:04.846 [2024-10-17 17:35:12.119858] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@864 -- # return 0 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@730 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 [2024-10-17 17:35:12.853018] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@21 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 Malloc0 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 Malloc1 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME -d SPDK_Controller1 -i 291 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 [2024-10-17 17:35:12.966025] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@28 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.106 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@30 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -a 10.0.0.2 -s 4420 00:14:05.366 00:14:05.366 Discovery Log Number of Records 2, Generation counter 2 00:14:05.366 =====Discovery Log Entry 0====== 00:14:05.366 trtype: tcp 00:14:05.366 adrfam: ipv4 00:14:05.366 subtype: current discovery subsystem 00:14:05.366 treq: not required 00:14:05.366 portid: 0 00:14:05.366 trsvcid: 4420 00:14:05.366 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:14:05.366 traddr: 10.0.0.2 00:14:05.366 eflags: explicit discovery connections, duplicate discovery information 00:14:05.366 sectype: none 00:14:05.366 =====Discovery Log Entry 1====== 00:14:05.366 trtype: tcp 00:14:05.366 adrfam: ipv4 00:14:05.366 subtype: nvme subsystem 00:14:05.366 treq: not required 00:14:05.366 portid: 0 00:14:05.366 trsvcid: 4420 00:14:05.366 subnqn: nqn.2016-06.io.spdk:cnode1 00:14:05.366 traddr: 10.0.0.2 00:14:05.366 eflags: none 00:14:05.366 sectype: none 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # devs=($(get_nvme_devs)) 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # get_nvme_devs 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@548 -- # local dev _ 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@547 -- # nvme list 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ Node == /dev/nvme* ]] 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ --------------------- == /dev/nvme* ]] 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # nvme_num_before_connection=0 00:14:05.366 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@32 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:14:06.750 17:35:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@34 -- # waitforserial SPDKISFASTANDAWESOME 2 00:14:06.750 17:35:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1198 -- # local i=0 00:14:06.750 17:35:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:14:06.750 17:35:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1200 -- # [[ -n 2 ]] 00:14:06.750 17:35:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1201 -- # nvme_device_counter=2 00:14:06.750 17:35:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1205 -- # sleep 2 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # nvme_devices=2 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1208 -- # return 0 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # get_nvme_devs 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@548 -- # local dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@547 -- # nvme list 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ Node == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ --------------------- == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n1 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n2 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # [[ -z /dev/nvme0n1 00:14:09.290 /dev/nvme0n2 ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # devs=($(get_nvme_devs)) 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # get_nvme_devs 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@548 -- # local dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@547 -- # nvme list 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ Node == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ --------------------- == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n1 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n2 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # nvme_num=2 00:14:09.290 17:35:16 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@60 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:14:09.549 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:14:09.549 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@61 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:14:09.549 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1219 -- # local i=0 00:14:09.549 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:14:09.549 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:09.549 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:14:09.549 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1231 -- # return 0 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@62 -- # (( nvme_num <= nvme_num_before_connection )) 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@67 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@70 -- # nvmftestfini 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@514 -- # nvmfcleanup 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@121 -- # sync 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@124 -- # set +e 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:09.550 rmmod nvme_tcp 00:14:09.550 rmmod nvme_fabrics 00:14:09.550 rmmod nvme_keyring 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@128 -- # set -e 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@129 -- # return 0 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@515 -- # '[' -n 4189069 ']' 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@516 -- # killprocess 4189069 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@950 -- # '[' -z 4189069 ']' 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@954 -- # kill -0 4189069 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@955 -- # uname 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4189069 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4189069' 00:14:09.550 killing process with pid 4189069 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@969 -- # kill 4189069 00:14:09.550 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@974 -- # wait 4189069 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@297 -- # iptr 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@789 -- # iptables-save 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@789 -- # iptables-restore 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:09.809 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:14:12.352 00:14:12.352 real 0m15.555s 00:14:12.352 user 0m24.005s 00:14:12.352 sys 0m6.428s 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:14:12.352 ************************************ 00:14:12.352 END TEST nvmf_nvme_cli 00:14:12.352 ************************************ 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@30 -- # [[ 1 -eq 1 ]] 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@31 -- # run_test nvmf_vfio_user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:14:12.352 ************************************ 00:14:12.352 START TEST nvmf_vfio_user 00:14:12.352 ************************************ 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:14:12.352 * Looking for test storage... 00:14:12.352 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1691 -- # lcov --version 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # IFS=.-: 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # read -ra ver1 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # IFS=.-: 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # read -ra ver2 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@338 -- # local 'op=<' 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@340 -- # ver1_l=2 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@341 -- # ver2_l=1 00:14:12.352 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@344 -- # case "$op" in 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@345 -- # : 1 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # decimal 1 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=1 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 1 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # ver1[v]=1 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # decimal 2 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=2 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 2 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # ver2[v]=2 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # return 0 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:14:12.353 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:12.353 --rc genhtml_branch_coverage=1 00:14:12.353 --rc genhtml_function_coverage=1 00:14:12.353 --rc genhtml_legend=1 00:14:12.353 --rc geninfo_all_blocks=1 00:14:12.353 --rc geninfo_unexecuted_blocks=1 00:14:12.353 00:14:12.353 ' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:14:12.353 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:12.353 --rc genhtml_branch_coverage=1 00:14:12.353 --rc genhtml_function_coverage=1 00:14:12.353 --rc genhtml_legend=1 00:14:12.353 --rc geninfo_all_blocks=1 00:14:12.353 --rc geninfo_unexecuted_blocks=1 00:14:12.353 00:14:12.353 ' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:14:12.353 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:12.353 --rc genhtml_branch_coverage=1 00:14:12.353 --rc genhtml_function_coverage=1 00:14:12.353 --rc genhtml_legend=1 00:14:12.353 --rc geninfo_all_blocks=1 00:14:12.353 --rc geninfo_unexecuted_blocks=1 00:14:12.353 00:14:12.353 ' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:14:12.353 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:12.353 --rc genhtml_branch_coverage=1 00:14:12.353 --rc genhtml_function_coverage=1 00:14:12.353 --rc genhtml_legend=1 00:14:12.353 --rc geninfo_all_blocks=1 00:14:12.353 --rc geninfo_unexecuted_blocks=1 00:14:12.353 00:14:12.353 ' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # uname -s 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@15 -- # shopt -s extglob 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@5 -- # export PATH 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@51 -- # : 0 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:12.353 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@12 -- # MALLOC_BDEV_SIZE=64 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@14 -- # NUM_DEVICES=2 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:12.353 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@47 -- # rm -rf /var/run/vfio-user 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@103 -- # setup_nvmf_vfio_user '' '' 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args= 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local transport_args= 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=4190770 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 4190770' 00:14:12.354 Process pid: 4190770 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 4190770 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@831 -- # '[' -z 4190770 ']' 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:12.354 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:12.354 17:35:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:14:12.354 [2024-10-17 17:35:20.027967] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:14:12.354 [2024-10-17 17:35:20.028053] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:12.354 [2024-10-17 17:35:20.114415] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:12.354 [2024-10-17 17:35:20.159543] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:12.354 [2024-10-17 17:35:20.159589] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:12.354 [2024-10-17 17:35:20.159596] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:12.354 [2024-10-17 17:35:20.159601] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:12.354 [2024-10-17 17:35:20.159605] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:12.354 [2024-10-17 17:35:20.161403] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:12.354 [2024-10-17 17:35:20.161557] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:12.354 [2024-10-17 17:35:20.161738] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:12.354 [2024-10-17 17:35:20.161761] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:12.925 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:12.925 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # return 0 00:14:12.925 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:14:14.308 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER 00:14:14.308 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:14:14.308 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:14:14.308 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:14:14.308 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:14:14.308 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:14:14.308 Malloc1 00:14:14.569 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:14:14.569 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:14:14.830 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:14:15.092 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:14:15.092 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:14:15.092 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:14:15.092 Malloc2 00:14:15.092 17:35:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:14:15.352 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:14:15.612 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:14:15.876 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@104 -- # run_nvmf_vfio_user 00:14:15.876 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # seq 1 2 00:14:15.876 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:14:15.876 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user1/1 00:14:15.876 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode1 00:14:15.876 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -L nvme -L nvme_vfio -L vfio_pci 00:14:15.876 [2024-10-17 17:35:23.571289] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:14:15.876 [2024-10-17 17:35:23.571360] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid4191469 ] 00:14:15.876 [2024-10-17 17:35:23.597290] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user1/1 00:14:15.876 [2024-10-17 17:35:23.609955] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:14:15.876 [2024-10-17 17:35:23.609972] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7fac96f46000 00:14:15.876 [2024-10-17 17:35:23.610958] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.611962] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.612963] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.613968] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.614972] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.615973] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.616983] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.617980] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:15.876 [2024-10-17 17:35:23.618988] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:14:15.876 [2024-10-17 17:35:23.618995] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7fac96f3b000 00:14:15.876 [2024-10-17 17:35:23.619909] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:14:15.876 [2024-10-17 17:35:23.629354] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user1/1/cntrl Setup Successfully 00:14:15.876 [2024-10-17 17:35:23.629376] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to connect adminq (no timeout) 00:14:15.876 [2024-10-17 17:35:23.634081] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:14:15.876 [2024-10-17 17:35:23.634114] nvme_pcie_common.c: 134:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:14:15.876 [2024-10-17 17:35:23.634174] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for connect adminq (no timeout) 00:14:15.876 [2024-10-17 17:35:23.634188] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read vs (no timeout) 00:14:15.876 [2024-10-17 17:35:23.634192] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read vs wait for vs (no timeout) 00:14:15.876 [2024-10-17 17:35:23.635079] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x8, value 0x10300 00:14:15.876 [2024-10-17 17:35:23.635086] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read cap (no timeout) 00:14:15.876 [2024-10-17 17:35:23.635091] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read cap wait for cap (no timeout) 00:14:15.876 [2024-10-17 17:35:23.636083] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:14:15.876 [2024-10-17 17:35:23.636090] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to check en (no timeout) 00:14:15.876 [2024-10-17 17:35:23.636098] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to check en wait for cc (timeout 15000 ms) 00:14:15.876 [2024-10-17 17:35:23.637094] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x0 00:14:15.876 [2024-10-17 17:35:23.637099] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:14:15.876 [2024-10-17 17:35:23.638098] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x0 00:14:15.876 [2024-10-17 17:35:23.638104] nvme_ctrlr.c:3893:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CC.EN = 0 && CSTS.RDY = 0 00:14:15.876 [2024-10-17 17:35:23.638107] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to controller is disabled (timeout 15000 ms) 00:14:15.876 [2024-10-17 17:35:23.638112] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:14:15.876 [2024-10-17 17:35:23.638216] nvme_ctrlr.c:4091:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Setting CC.EN = 1 00:14:15.876 [2024-10-17 17:35:23.638219] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:14:15.876 [2024-10-17 17:35:23.638223] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x28, value 0x2000003c0000 00:14:15.876 [2024-10-17 17:35:23.639105] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x30, value 0x2000003be000 00:14:15.876 [2024-10-17 17:35:23.640105] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x24, value 0xff00ff 00:14:15.876 [2024-10-17 17:35:23.641115] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:14:15.876 [2024-10-17 17:35:23.642107] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:15.876 [2024-10-17 17:35:23.642175] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:14:15.876 [2024-10-17 17:35:23.643125] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x1 00:14:15.876 [2024-10-17 17:35:23.643131] nvme_ctrlr.c:3928:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:14:15.876 [2024-10-17 17:35:23.643135] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to reset admin queue (timeout 30000 ms) 00:14:15.876 [2024-10-17 17:35:23.643149] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify controller (no timeout) 00:14:15.876 [2024-10-17 17:35:23.643157] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify controller (timeout 30000 ms) 00:14:15.876 [2024-10-17 17:35:23.643168] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:15.876 [2024-10-17 17:35:23.643171] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:15.876 [2024-10-17 17:35:23.643174] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:15.876 [2024-10-17 17:35:23.643184] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:15.876 [2024-10-17 17:35:23.643227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:14:15.876 [2024-10-17 17:35:23.643236] nvme_ctrlr.c:2077:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] transport max_xfer_size 131072 00:14:15.876 [2024-10-17 17:35:23.643239] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] MDTS max_xfer_size 131072 00:14:15.876 [2024-10-17 17:35:23.643242] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CNTLID 0x0001 00:14:15.876 [2024-10-17 17:35:23.643246] nvme_ctrlr.c:2095:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:14:15.876 [2024-10-17 17:35:23.643249] nvme_ctrlr.c:2108:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] transport max_sges 1 00:14:15.876 [2024-10-17 17:35:23.643252] nvme_ctrlr.c:2123:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] fuses compare and write: 1 00:14:15.876 [2024-10-17 17:35:23.643256] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to configure AER (timeout 30000 ms) 00:14:15.876 [2024-10-17 17:35:23.643262] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for configure aer (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643269] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643291] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:14:15.877 [2024-10-17 17:35:23.643297] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:14:15.877 [2024-10-17 17:35:23.643303] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:14:15.877 [2024-10-17 17:35:23.643309] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:14:15.877 [2024-10-17 17:35:23.643312] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set keep alive timeout (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643319] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643325] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643340] nvme_ctrlr.c:3034:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Controller adjusted keep alive timeout to 0 ms 00:14:15.877 [2024-10-17 17:35:23.643344] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify controller iocs specific (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643348] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set number of queues (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643354] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for set number of queues (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643360] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643414] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify active ns (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643421] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify active ns (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643427] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:14:15.877 [2024-10-17 17:35:23.643430] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:14:15.877 [2024-10-17 17:35:23.643432] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:15.877 [2024-10-17 17:35:23.643437] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643453] nvme_ctrlr.c:4722:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Namespace 1 was added 00:14:15.877 [2024-10-17 17:35:23.643459] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify ns (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643464] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify ns (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643469] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:15.877 [2024-10-17 17:35:23.643472] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:15.877 [2024-10-17 17:35:23.643475] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:15.877 [2024-10-17 17:35:23.643479] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643505] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643511] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643516] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:15.877 [2024-10-17 17:35:23.643519] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:15.877 [2024-10-17 17:35:23.643521] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:15.877 [2024-10-17 17:35:23.643526] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643544] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify ns iocs specific (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643548] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set supported log pages (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643554] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set supported features (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643558] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set host behavior support feature (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643561] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set doorbell buffer config (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643566] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set host ID (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643570] nvme_ctrlr.c:3134:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] NVMe-oF transport - not sending Set Features - Host ID 00:14:15.877 [2024-10-17 17:35:23.643573] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to transport ready (timeout 30000 ms) 00:14:15.877 [2024-10-17 17:35:23.643577] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to ready (no timeout) 00:14:15.877 [2024-10-17 17:35:23.643590] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643608] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643626] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643646] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643662] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:14:15.877 [2024-10-17 17:35:23.643665] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:14:15.877 [2024-10-17 17:35:23.643668] nvme_pcie_common.c:1241:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:14:15.877 [2024-10-17 17:35:23.643670] nvme_pcie_common.c:1257:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:14:15.877 [2024-10-17 17:35:23.643673] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:14:15.877 [2024-10-17 17:35:23.643677] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:14:15.877 [2024-10-17 17:35:23.643682] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:14:15.877 [2024-10-17 17:35:23.643685] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:14:15.877 [2024-10-17 17:35:23.643688] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:15.877 [2024-10-17 17:35:23.643695] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643701] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:14:15.877 [2024-10-17 17:35:23.643704] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:15.877 [2024-10-17 17:35:23.643706] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:15.877 [2024-10-17 17:35:23.643710] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643716] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:14:15.877 [2024-10-17 17:35:23.643719] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:14:15.877 [2024-10-17 17:35:23.643722] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:15.877 [2024-10-17 17:35:23.643727] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:14:15.877 [2024-10-17 17:35:23.643732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:14:15.877 [2024-10-17 17:35:23.643753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:14:15.877 ===================================================== 00:14:15.877 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:14:15.877 ===================================================== 00:14:15.877 Controller Capabilities/Features 00:14:15.877 ================================ 00:14:15.877 Vendor ID: 4e58 00:14:15.877 Subsystem Vendor ID: 4e58 00:14:15.877 Serial Number: SPDK1 00:14:15.877 Model Number: SPDK bdev Controller 00:14:15.877 Firmware Version: 25.01 00:14:15.877 Recommended Arb Burst: 6 00:14:15.877 IEEE OUI Identifier: 8d 6b 50 00:14:15.877 Multi-path I/O 00:14:15.877 May have multiple subsystem ports: Yes 00:14:15.877 May have multiple controllers: Yes 00:14:15.877 Associated with SR-IOV VF: No 00:14:15.877 Max Data Transfer Size: 131072 00:14:15.877 Max Number of Namespaces: 32 00:14:15.877 Max Number of I/O Queues: 127 00:14:15.877 NVMe Specification Version (VS): 1.3 00:14:15.877 NVMe Specification Version (Identify): 1.3 00:14:15.877 Maximum Queue Entries: 256 00:14:15.877 Contiguous Queues Required: Yes 00:14:15.877 Arbitration Mechanisms Supported 00:14:15.877 Weighted Round Robin: Not Supported 00:14:15.878 Vendor Specific: Not Supported 00:14:15.878 Reset Timeout: 15000 ms 00:14:15.878 Doorbell Stride: 4 bytes 00:14:15.878 NVM Subsystem Reset: Not Supported 00:14:15.878 Command Sets Supported 00:14:15.878 NVM Command Set: Supported 00:14:15.878 Boot Partition: Not Supported 00:14:15.878 Memory Page Size Minimum: 4096 bytes 00:14:15.878 Memory Page Size Maximum: 4096 bytes 00:14:15.878 Persistent Memory Region: Not Supported 00:14:15.878 Optional Asynchronous Events Supported 00:14:15.878 Namespace Attribute Notices: Supported 00:14:15.878 Firmware Activation Notices: Not Supported 00:14:15.878 ANA Change Notices: Not Supported 00:14:15.878 PLE Aggregate Log Change Notices: Not Supported 00:14:15.878 LBA Status Info Alert Notices: Not Supported 00:14:15.878 EGE Aggregate Log Change Notices: Not Supported 00:14:15.878 Normal NVM Subsystem Shutdown event: Not Supported 00:14:15.878 Zone Descriptor Change Notices: Not Supported 00:14:15.878 Discovery Log Change Notices: Not Supported 00:14:15.878 Controller Attributes 00:14:15.878 128-bit Host Identifier: Supported 00:14:15.878 Non-Operational Permissive Mode: Not Supported 00:14:15.878 NVM Sets: Not Supported 00:14:15.878 Read Recovery Levels: Not Supported 00:14:15.878 Endurance Groups: Not Supported 00:14:15.878 Predictable Latency Mode: Not Supported 00:14:15.878 Traffic Based Keep ALive: Not Supported 00:14:15.878 Namespace Granularity: Not Supported 00:14:15.878 SQ Associations: Not Supported 00:14:15.878 UUID List: Not Supported 00:14:15.878 Multi-Domain Subsystem: Not Supported 00:14:15.878 Fixed Capacity Management: Not Supported 00:14:15.878 Variable Capacity Management: Not Supported 00:14:15.878 Delete Endurance Group: Not Supported 00:14:15.878 Delete NVM Set: Not Supported 00:14:15.878 Extended LBA Formats Supported: Not Supported 00:14:15.878 Flexible Data Placement Supported: Not Supported 00:14:15.878 00:14:15.878 Controller Memory Buffer Support 00:14:15.878 ================================ 00:14:15.878 Supported: No 00:14:15.878 00:14:15.878 Persistent Memory Region Support 00:14:15.878 ================================ 00:14:15.878 Supported: No 00:14:15.878 00:14:15.878 Admin Command Set Attributes 00:14:15.878 ============================ 00:14:15.878 Security Send/Receive: Not Supported 00:14:15.878 Format NVM: Not Supported 00:14:15.878 Firmware Activate/Download: Not Supported 00:14:15.878 Namespace Management: Not Supported 00:14:15.878 Device Self-Test: Not Supported 00:14:15.878 Directives: Not Supported 00:14:15.878 NVMe-MI: Not Supported 00:14:15.878 Virtualization Management: Not Supported 00:14:15.878 Doorbell Buffer Config: Not Supported 00:14:15.878 Get LBA Status Capability: Not Supported 00:14:15.878 Command & Feature Lockdown Capability: Not Supported 00:14:15.878 Abort Command Limit: 4 00:14:15.878 Async Event Request Limit: 4 00:14:15.878 Number of Firmware Slots: N/A 00:14:15.878 Firmware Slot 1 Read-Only: N/A 00:14:15.878 Firmware Activation Without Reset: N/A 00:14:15.878 Multiple Update Detection Support: N/A 00:14:15.878 Firmware Update Granularity: No Information Provided 00:14:15.878 Per-Namespace SMART Log: No 00:14:15.878 Asymmetric Namespace Access Log Page: Not Supported 00:14:15.878 Subsystem NQN: nqn.2019-07.io.spdk:cnode1 00:14:15.878 Command Effects Log Page: Supported 00:14:15.878 Get Log Page Extended Data: Supported 00:14:15.878 Telemetry Log Pages: Not Supported 00:14:15.878 Persistent Event Log Pages: Not Supported 00:14:15.878 Supported Log Pages Log Page: May Support 00:14:15.878 Commands Supported & Effects Log Page: Not Supported 00:14:15.878 Feature Identifiers & Effects Log Page:May Support 00:14:15.878 NVMe-MI Commands & Effects Log Page: May Support 00:14:15.878 Data Area 4 for Telemetry Log: Not Supported 00:14:15.878 Error Log Page Entries Supported: 128 00:14:15.878 Keep Alive: Supported 00:14:15.878 Keep Alive Granularity: 10000 ms 00:14:15.878 00:14:15.878 NVM Command Set Attributes 00:14:15.878 ========================== 00:14:15.878 Submission Queue Entry Size 00:14:15.878 Max: 64 00:14:15.878 Min: 64 00:14:15.878 Completion Queue Entry Size 00:14:15.878 Max: 16 00:14:15.878 Min: 16 00:14:15.878 Number of Namespaces: 32 00:14:15.878 Compare Command: Supported 00:14:15.878 Write Uncorrectable Command: Not Supported 00:14:15.878 Dataset Management Command: Supported 00:14:15.878 Write Zeroes Command: Supported 00:14:15.878 Set Features Save Field: Not Supported 00:14:15.878 Reservations: Not Supported 00:14:15.878 Timestamp: Not Supported 00:14:15.878 Copy: Supported 00:14:15.878 Volatile Write Cache: Present 00:14:15.878 Atomic Write Unit (Normal): 1 00:14:15.878 Atomic Write Unit (PFail): 1 00:14:15.878 Atomic Compare & Write Unit: 1 00:14:15.878 Fused Compare & Write: Supported 00:14:15.878 Scatter-Gather List 00:14:15.878 SGL Command Set: Supported (Dword aligned) 00:14:15.878 SGL Keyed: Not Supported 00:14:15.878 SGL Bit Bucket Descriptor: Not Supported 00:14:15.878 SGL Metadata Pointer: Not Supported 00:14:15.878 Oversized SGL: Not Supported 00:14:15.878 SGL Metadata Address: Not Supported 00:14:15.878 SGL Offset: Not Supported 00:14:15.878 Transport SGL Data Block: Not Supported 00:14:15.878 Replay Protected Memory Block: Not Supported 00:14:15.878 00:14:15.878 Firmware Slot Information 00:14:15.878 ========================= 00:14:15.878 Active slot: 1 00:14:15.878 Slot 1 Firmware Revision: 25.01 00:14:15.878 00:14:15.878 00:14:15.878 Commands Supported and Effects 00:14:15.878 ============================== 00:14:15.878 Admin Commands 00:14:15.878 -------------- 00:14:15.878 Get Log Page (02h): Supported 00:14:15.878 Identify (06h): Supported 00:14:15.878 Abort (08h): Supported 00:14:15.878 Set Features (09h): Supported 00:14:15.878 Get Features (0Ah): Supported 00:14:15.878 Asynchronous Event Request (0Ch): Supported 00:14:15.878 Keep Alive (18h): Supported 00:14:15.878 I/O Commands 00:14:15.878 ------------ 00:14:15.878 Flush (00h): Supported LBA-Change 00:14:15.878 Write (01h): Supported LBA-Change 00:14:15.878 Read (02h): Supported 00:14:15.878 Compare (05h): Supported 00:14:15.878 Write Zeroes (08h): Supported LBA-Change 00:14:15.878 Dataset Management (09h): Supported LBA-Change 00:14:15.878 Copy (19h): Supported LBA-Change 00:14:15.878 00:14:15.878 Error Log 00:14:15.878 ========= 00:14:15.878 00:14:15.878 Arbitration 00:14:15.878 =========== 00:14:15.878 Arbitration Burst: 1 00:14:15.878 00:14:15.878 Power Management 00:14:15.878 ================ 00:14:15.878 Number of Power States: 1 00:14:15.878 Current Power State: Power State #0 00:14:15.878 Power State #0: 00:14:15.878 Max Power: 0.00 W 00:14:15.878 Non-Operational State: Operational 00:14:15.878 Entry Latency: Not Reported 00:14:15.878 Exit Latency: Not Reported 00:14:15.878 Relative Read Throughput: 0 00:14:15.878 Relative Read Latency: 0 00:14:15.878 Relative Write Throughput: 0 00:14:15.878 Relative Write Latency: 0 00:14:15.878 Idle Power: Not Reported 00:14:15.878 Active Power: Not Reported 00:14:15.878 Non-Operational Permissive Mode: Not Supported 00:14:15.878 00:14:15.878 Health Information 00:14:15.878 ================== 00:14:15.878 Critical Warnings: 00:14:15.878 Available Spare Space: OK 00:14:15.878 Temperature: OK 00:14:15.878 Device Reliability: OK 00:14:15.878 Read Only: No 00:14:15.878 Volatile Memory Backup: OK 00:14:15.878 Current Temperature: 0 Kelvin (-273 Celsius) 00:14:15.878 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:14:15.878 Available Spare: 0% 00:14:15.878 Available Sp[2024-10-17 17:35:23.643828] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:14:15.878 [2024-10-17 17:35:23.643839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:14:15.878 [2024-10-17 17:35:23.643860] nvme_ctrlr.c:4386:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Prepare to destruct SSD 00:14:15.878 [2024-10-17 17:35:23.643867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:15.878 [2024-10-17 17:35:23.643872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:15.878 [2024-10-17 17:35:23.643876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:15.878 [2024-10-17 17:35:23.643880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:15.878 [2024-10-17 17:35:23.644129] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:14:15.878 [2024-10-17 17:35:23.644136] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x464001 00:14:15.878 [2024-10-17 17:35:23.645134] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:15.878 [2024-10-17 17:35:23.645172] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] RTD3E = 0 us 00:14:15.878 [2024-10-17 17:35:23.645176] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] shutdown timeout = 10000 ms 00:14:15.878 [2024-10-17 17:35:23.646146] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x9 00:14:15.878 [2024-10-17 17:35:23.646154] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] shutdown complete in 0 milliseconds 00:14:15.879 [2024-10-17 17:35:23.646213] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user1/1/cntrl 00:14:15.879 [2024-10-17 17:35:23.648698] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:14:15.879 are Threshold: 0% 00:14:15.879 Life Percentage Used: 0% 00:14:15.879 Data Units Read: 0 00:14:15.879 Data Units Written: 0 00:14:15.879 Host Read Commands: 0 00:14:15.879 Host Write Commands: 0 00:14:15.879 Controller Busy Time: 0 minutes 00:14:15.879 Power Cycles: 0 00:14:15.879 Power On Hours: 0 hours 00:14:15.879 Unsafe Shutdowns: 0 00:14:15.879 Unrecoverable Media Errors: 0 00:14:15.879 Lifetime Error Log Entries: 0 00:14:15.879 Warning Temperature Time: 0 minutes 00:14:15.879 Critical Temperature Time: 0 minutes 00:14:15.879 00:14:15.879 Number of Queues 00:14:15.879 ================ 00:14:15.879 Number of I/O Submission Queues: 127 00:14:15.879 Number of I/O Completion Queues: 127 00:14:15.879 00:14:15.879 Active Namespaces 00:14:15.879 ================= 00:14:15.879 Namespace ID:1 00:14:15.879 Error Recovery Timeout: Unlimited 00:14:15.879 Command Set Identifier: NVM (00h) 00:14:15.879 Deallocate: Supported 00:14:15.879 Deallocated/Unwritten Error: Not Supported 00:14:15.879 Deallocated Read Value: Unknown 00:14:15.879 Deallocate in Write Zeroes: Not Supported 00:14:15.879 Deallocated Guard Field: 0xFFFF 00:14:15.879 Flush: Supported 00:14:15.879 Reservation: Supported 00:14:15.879 Namespace Sharing Capabilities: Multiple Controllers 00:14:15.879 Size (in LBAs): 131072 (0GiB) 00:14:15.879 Capacity (in LBAs): 131072 (0GiB) 00:14:15.879 Utilization (in LBAs): 131072 (0GiB) 00:14:15.879 NGUID: 98E788A861DA4B87A0F94BD5380D6C9A 00:14:15.879 UUID: 98e788a8-61da-4b87-a0f9-4bd5380d6c9a 00:14:15.879 Thin Provisioning: Not Supported 00:14:15.879 Per-NS Atomic Units: Yes 00:14:15.879 Atomic Boundary Size (Normal): 0 00:14:15.879 Atomic Boundary Size (PFail): 0 00:14:15.879 Atomic Boundary Offset: 0 00:14:15.879 Maximum Single Source Range Length: 65535 00:14:15.879 Maximum Copy Length: 65535 00:14:15.879 Maximum Source Range Count: 1 00:14:15.879 NGUID/EUI64 Never Reused: No 00:14:15.879 Namespace Write Protected: No 00:14:15.879 Number of LBA Formats: 1 00:14:15.879 Current LBA Format: LBA Format #00 00:14:15.879 LBA Format #00: Data Size: 512 Metadata Size: 0 00:14:15.879 00:14:15.879 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:14:16.139 [2024-10-17 17:35:23.817345] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:21.505 Initializing NVMe Controllers 00:14:21.505 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:14:21.505 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:14:21.505 Initialization complete. Launching workers. 00:14:21.505 ======================================================== 00:14:21.505 Latency(us) 00:14:21.505 Device Information : IOPS MiB/s Average min max 00:14:21.505 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 39995.40 156.23 3200.59 851.37 7195.56 00:14:21.505 ======================================================== 00:14:21.506 Total : 39995.40 156.23 3200.59 851.37 7195.56 00:14:21.506 00:14:21.506 [2024-10-17 17:35:28.834497] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:21.506 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:14:21.506 [2024-10-17 17:35:29.007298] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:26.789 Initializing NVMe Controllers 00:14:26.789 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:14:26.789 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:14:26.789 Initialization complete. Launching workers. 00:14:26.789 ======================================================== 00:14:26.789 Latency(us) 00:14:26.789 Device Information : IOPS MiB/s Average min max 00:14:26.789 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 16051.20 62.70 7982.32 5985.48 11973.03 00:14:26.789 ======================================================== 00:14:26.789 Total : 16051.20 62.70 7982.32 5985.48 11973.03 00:14:26.789 00:14:26.789 [2024-10-17 17:35:34.040207] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:26.789 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:14:26.789 [2024-10-17 17:35:34.228995] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:32.070 [2024-10-17 17:35:39.312912] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:32.070 Initializing NVMe Controllers 00:14:32.070 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:14:32.070 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:14:32.070 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 1 00:14:32.070 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 2 00:14:32.070 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 3 00:14:32.070 Initialization complete. Launching workers. 00:14:32.070 Starting thread on core 2 00:14:32.070 Starting thread on core 3 00:14:32.070 Starting thread on core 1 00:14:32.070 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -d 256 -g 00:14:32.070 [2024-10-17 17:35:39.547165] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:35.366 [2024-10-17 17:35:42.611696] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:35.366 Initializing NVMe Controllers 00:14:35.366 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:35.366 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:35.366 Associating SPDK bdev Controller (SPDK1 ) with lcore 0 00:14:35.367 Associating SPDK bdev Controller (SPDK1 ) with lcore 1 00:14:35.367 Associating SPDK bdev Controller (SPDK1 ) with lcore 2 00:14:35.367 Associating SPDK bdev Controller (SPDK1 ) with lcore 3 00:14:35.367 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:14:35.367 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:14:35.367 Initialization complete. Launching workers. 00:14:35.367 Starting thread on core 1 with urgent priority queue 00:14:35.367 Starting thread on core 2 with urgent priority queue 00:14:35.367 Starting thread on core 3 with urgent priority queue 00:14:35.367 Starting thread on core 0 with urgent priority queue 00:14:35.367 SPDK bdev Controller (SPDK1 ) core 0: 8520.00 IO/s 11.74 secs/100000 ios 00:14:35.367 SPDK bdev Controller (SPDK1 ) core 1: 10419.67 IO/s 9.60 secs/100000 ios 00:14:35.367 SPDK bdev Controller (SPDK1 ) core 2: 10141.33 IO/s 9.86 secs/100000 ios 00:14:35.367 SPDK bdev Controller (SPDK1 ) core 3: 11326.67 IO/s 8.83 secs/100000 ios 00:14:35.367 ======================================================== 00:14:35.367 00:14:35.367 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:14:35.367 [2024-10-17 17:35:42.835304] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:35.367 Initializing NVMe Controllers 00:14:35.367 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:35.367 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:35.367 Namespace ID: 1 size: 0GB 00:14:35.367 Initialization complete. 00:14:35.367 INFO: using host memory buffer for IO 00:14:35.367 Hello world! 00:14:35.367 [2024-10-17 17:35:42.869518] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:35.367 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:14:35.367 [2024-10-17 17:35:43.099524] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:36.309 Initializing NVMe Controllers 00:14:36.309 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:36.309 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:36.309 Initialization complete. Launching workers. 00:14:36.310 submit (in ns) avg, min, max = 5928.2, 2836.7, 3998105.0 00:14:36.310 complete (in ns) avg, min, max = 16798.1, 1635.0, 3997152.5 00:14:36.310 00:14:36.310 Submit histogram 00:14:36.310 ================ 00:14:36.310 Range in us Cumulative Count 00:14:36.310 2.827 - 2.840: 0.0295% ( 6) 00:14:36.310 2.840 - 2.853: 0.5400% ( 104) 00:14:36.310 2.853 - 2.867: 2.1501% ( 328) 00:14:36.310 2.867 - 2.880: 4.7666% ( 533) 00:14:36.310 2.880 - 2.893: 9.5184% ( 968) 00:14:36.310 2.893 - 2.907: 15.4975% ( 1218) 00:14:36.310 2.907 - 2.920: 21.9086% ( 1306) 00:14:36.310 2.920 - 2.933: 27.7404% ( 1188) 00:14:36.310 2.933 - 2.947: 32.9832% ( 1068) 00:14:36.310 2.947 - 2.960: 38.7708% ( 1179) 00:14:36.310 2.960 - 2.973: 45.0935% ( 1288) 00:14:36.310 2.973 - 2.987: 52.2802% ( 1464) 00:14:36.310 2.987 - 3.000: 59.8940% ( 1551) 00:14:36.310 3.000 - 3.013: 67.8170% ( 1614) 00:14:36.310 3.013 - 3.027: 75.6173% ( 1589) 00:14:36.310 3.027 - 3.040: 82.9218% ( 1488) 00:14:36.310 3.040 - 3.053: 89.3132% ( 1302) 00:14:36.310 3.053 - 3.067: 93.9522% ( 945) 00:14:36.310 3.067 - 3.080: 97.0105% ( 623) 00:14:36.310 3.080 - 3.093: 98.6451% ( 333) 00:14:36.310 3.093 - 3.107: 99.2588% ( 125) 00:14:36.310 3.107 - 3.120: 99.4600% ( 41) 00:14:36.310 3.120 - 3.133: 99.4895% ( 6) 00:14:36.310 3.133 - 3.147: 99.5091% ( 4) 00:14:36.310 3.147 - 3.160: 99.5189% ( 2) 00:14:36.310 3.160 - 3.173: 99.5337% ( 3) 00:14:36.310 3.227 - 3.240: 99.5386% ( 1) 00:14:36.310 3.267 - 3.280: 99.5435% ( 1) 00:14:36.310 3.307 - 3.320: 99.5484% ( 1) 00:14:36.310 3.467 - 3.493: 99.5533% ( 1) 00:14:36.310 3.493 - 3.520: 99.5582% ( 1) 00:14:36.310 3.573 - 3.600: 99.5631% ( 1) 00:14:36.310 3.600 - 3.627: 99.5680% ( 1) 00:14:36.310 3.627 - 3.653: 99.5729% ( 1) 00:14:36.310 3.733 - 3.760: 99.5778% ( 1) 00:14:36.310 3.787 - 3.813: 99.5827% ( 1) 00:14:36.310 4.027 - 4.053: 99.5926% ( 2) 00:14:36.310 4.187 - 4.213: 99.5975% ( 1) 00:14:36.310 4.587 - 4.613: 99.6024% ( 1) 00:14:36.310 4.667 - 4.693: 99.6073% ( 1) 00:14:36.310 4.693 - 4.720: 99.6122% ( 1) 00:14:36.310 4.747 - 4.773: 99.6171% ( 1) 00:14:36.310 4.773 - 4.800: 99.6318% ( 3) 00:14:36.310 4.800 - 4.827: 99.6367% ( 1) 00:14:36.310 4.827 - 4.853: 99.6416% ( 1) 00:14:36.310 4.853 - 4.880: 99.6515% ( 2) 00:14:36.310 4.880 - 4.907: 99.6564% ( 1) 00:14:36.310 4.907 - 4.933: 99.6662% ( 2) 00:14:36.310 4.960 - 4.987: 99.6711% ( 1) 00:14:36.310 4.987 - 5.013: 99.6760% ( 1) 00:14:36.310 5.013 - 5.040: 99.6809% ( 1) 00:14:36.310 5.040 - 5.067: 99.6907% ( 2) 00:14:36.310 5.067 - 5.093: 99.6956% ( 1) 00:14:36.310 5.093 - 5.120: 99.7006% ( 1) 00:14:36.310 5.147 - 5.173: 99.7055% ( 1) 00:14:36.310 5.173 - 5.200: 99.7104% ( 1) 00:14:36.310 5.253 - 5.280: 99.7251% ( 3) 00:14:36.310 5.280 - 5.307: 99.7349% ( 2) 00:14:36.310 5.307 - 5.333: 99.7496% ( 3) 00:14:36.310 5.360 - 5.387: 99.7546% ( 1) 00:14:36.310 5.387 - 5.413: 99.7595% ( 1) 00:14:36.310 5.467 - 5.493: 99.7644% ( 1) 00:14:36.310 5.520 - 5.547: 99.7791% ( 3) 00:14:36.310 5.600 - 5.627: 99.7840% ( 1) 00:14:36.310 5.627 - 5.653: 99.7889% ( 1) 00:14:36.310 5.680 - 5.707: 99.7938% ( 1) 00:14:36.310 5.733 - 5.760: 99.7987% ( 1) 00:14:36.310 5.787 - 5.813: 99.8036% ( 1) 00:14:36.310 5.840 - 5.867: 99.8086% ( 1) 00:14:36.310 5.893 - 5.920: 99.8135% ( 1) 00:14:36.310 5.920 - 5.947: 99.8184% ( 1) 00:14:36.310 5.973 - 6.000: 99.8233% ( 1) 00:14:36.310 6.080 - 6.107: 99.8282% ( 1) 00:14:36.310 6.213 - 6.240: 99.8380% ( 2) 00:14:36.310 6.240 - 6.267: 99.8478% ( 2) 00:14:36.310 6.320 - 6.347: 99.8527% ( 1) 00:14:36.310 6.347 - 6.373: 99.8576% ( 1) 00:14:36.310 6.373 - 6.400: 99.8625% ( 1) 00:14:36.310 6.533 - 6.560: 99.8724% ( 2) 00:14:36.310 [2024-10-17 17:35:44.118091] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:36.310 6.587 - 6.613: 99.8773% ( 1) 00:14:36.310 6.667 - 6.693: 99.8822% ( 1) 00:14:36.310 6.773 - 6.800: 99.8871% ( 1) 00:14:36.310 6.933 - 6.987: 99.8920% ( 1) 00:14:36.310 7.040 - 7.093: 99.8969% ( 1) 00:14:36.310 7.093 - 7.147: 99.9067% ( 2) 00:14:36.310 7.573 - 7.627: 99.9116% ( 1) 00:14:36.310 10.773 - 10.827: 99.9165% ( 1) 00:14:36.310 12.053 - 12.107: 99.9215% ( 1) 00:14:36.310 12.907 - 12.960: 99.9264% ( 1) 00:14:36.310 3986.773 - 4014.080: 100.0000% ( 15) 00:14:36.310 00:14:36.310 Complete histogram 00:14:36.310 ================== 00:14:36.310 Range in us Cumulative Count 00:14:36.310 1.633 - 1.640: 0.3289% ( 67) 00:14:36.310 1.640 - 1.647: 0.7560% ( 87) 00:14:36.310 1.647 - 1.653: 0.8002% ( 9) 00:14:36.310 1.653 - 1.660: 0.9622% ( 33) 00:14:36.310 1.660 - 1.667: 1.0505% ( 18) 00:14:36.310 1.667 - 1.673: 1.0800% ( 6) 00:14:36.310 1.673 - 1.680: 1.0898% ( 2) 00:14:36.310 1.680 - 1.687: 3.6326% ( 518) 00:14:36.310 1.687 - 1.693: 29.7433% ( 5319) 00:14:36.310 1.693 - 1.700: 37.0379% ( 1486) 00:14:36.310 1.700 - 1.707: 50.2675% ( 2695) 00:14:36.310 1.707 - 1.720: 74.5226% ( 4941) 00:14:36.310 1.720 - 1.733: 82.6567% ( 1657) 00:14:36.310 1.733 - 1.747: 84.3847% ( 352) 00:14:36.310 1.747 - 1.760: 87.8946% ( 715) 00:14:36.310 1.760 - 1.773: 93.2551% ( 1092) 00:14:36.310 1.773 - 1.787: 97.1823% ( 800) 00:14:36.310 1.787 - 1.800: 98.8464% ( 339) 00:14:36.310 1.800 - 1.813: 99.3324% ( 99) 00:14:36.310 1.813 - 1.827: 99.4600% ( 26) 00:14:36.310 1.960 - 1.973: 99.4649% ( 1) 00:14:36.310 2.067 - 2.080: 99.4698% ( 1) 00:14:36.310 2.280 - 2.293: 99.4747% ( 1) 00:14:36.310 3.600 - 3.627: 99.4797% ( 1) 00:14:36.310 3.680 - 3.707: 99.4846% ( 1) 00:14:36.310 3.707 - 3.733: 99.4944% ( 2) 00:14:36.310 3.733 - 3.760: 99.4993% ( 1) 00:14:36.310 3.760 - 3.787: 99.5042% ( 1) 00:14:36.310 3.867 - 3.893: 99.5091% ( 1) 00:14:36.310 4.027 - 4.053: 99.5140% ( 1) 00:14:36.310 4.107 - 4.133: 99.5189% ( 1) 00:14:36.310 4.133 - 4.160: 99.5287% ( 2) 00:14:36.310 4.320 - 4.347: 99.5337% ( 1) 00:14:36.310 4.347 - 4.373: 99.5435% ( 2) 00:14:36.310 4.400 - 4.427: 99.5484% ( 1) 00:14:36.310 4.453 - 4.480: 99.5533% ( 1) 00:14:36.310 4.480 - 4.507: 99.5582% ( 1) 00:14:36.310 4.507 - 4.533: 99.5729% ( 3) 00:14:36.310 4.800 - 4.827: 99.5778% ( 1) 00:14:36.310 4.853 - 4.880: 99.5876% ( 2) 00:14:36.310 4.880 - 4.907: 99.5926% ( 1) 00:14:36.310 4.987 - 5.013: 99.5975% ( 1) 00:14:36.310 5.200 - 5.227: 99.6024% ( 1) 00:14:36.310 5.600 - 5.627: 99.6073% ( 1) 00:14:36.310 9.493 - 9.547: 99.6122% ( 1) 00:14:36.310 11.627 - 11.680: 99.6171% ( 1) 00:14:36.310 34.560 - 34.773: 99.6220% ( 1) 00:14:36.310 3604.480 - 3631.787: 99.6269% ( 1) 00:14:36.310 3986.773 - 4014.080: 100.0000% ( 76) 00:14:36.310 00:14:36.310 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user1/1 nqn.2019-07.io.spdk:cnode1 1 00:14:36.310 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user1/1 00:14:36.310 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode1 00:14:36.310 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc3 00:14:36.310 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:36.572 [ 00:14:36.572 { 00:14:36.572 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:36.572 "subtype": "Discovery", 00:14:36.572 "listen_addresses": [], 00:14:36.572 "allow_any_host": true, 00:14:36.572 "hosts": [] 00:14:36.572 }, 00:14:36.572 { 00:14:36.572 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:36.572 "subtype": "NVMe", 00:14:36.572 "listen_addresses": [ 00:14:36.572 { 00:14:36.572 "trtype": "VFIOUSER", 00:14:36.572 "adrfam": "IPv4", 00:14:36.572 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:36.572 "trsvcid": "0" 00:14:36.572 } 00:14:36.572 ], 00:14:36.572 "allow_any_host": true, 00:14:36.572 "hosts": [], 00:14:36.572 "serial_number": "SPDK1", 00:14:36.572 "model_number": "SPDK bdev Controller", 00:14:36.572 "max_namespaces": 32, 00:14:36.572 "min_cntlid": 1, 00:14:36.572 "max_cntlid": 65519, 00:14:36.572 "namespaces": [ 00:14:36.572 { 00:14:36.572 "nsid": 1, 00:14:36.572 "bdev_name": "Malloc1", 00:14:36.572 "name": "Malloc1", 00:14:36.572 "nguid": "98E788A861DA4B87A0F94BD5380D6C9A", 00:14:36.572 "uuid": "98e788a8-61da-4b87-a0f9-4bd5380d6c9a" 00:14:36.572 } 00:14:36.572 ] 00:14:36.572 }, 00:14:36.572 { 00:14:36.572 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:36.572 "subtype": "NVMe", 00:14:36.572 "listen_addresses": [ 00:14:36.572 { 00:14:36.572 "trtype": "VFIOUSER", 00:14:36.572 "adrfam": "IPv4", 00:14:36.572 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:36.572 "trsvcid": "0" 00:14:36.572 } 00:14:36.572 ], 00:14:36.572 "allow_any_host": true, 00:14:36.572 "hosts": [], 00:14:36.572 "serial_number": "SPDK2", 00:14:36.572 "model_number": "SPDK bdev Controller", 00:14:36.572 "max_namespaces": 32, 00:14:36.572 "min_cntlid": 1, 00:14:36.572 "max_cntlid": 65519, 00:14:36.572 "namespaces": [ 00:14:36.572 { 00:14:36.572 "nsid": 1, 00:14:36.572 "bdev_name": "Malloc2", 00:14:36.572 "name": "Malloc2", 00:14:36.572 "nguid": "6FA7B43C200F40C5BE19519108931ED7", 00:14:36.572 "uuid": "6fa7b43c-200f-40c5-be19-519108931ed7" 00:14:36.572 } 00:14:36.572 ] 00:14:36.572 } 00:14:36.572 ] 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -n 2 -g -t /tmp/aer_touch_file 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=2114 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1265 -- # local i=0 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # return 0 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:14:36.572 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc3 00:14:36.572 [2024-10-17 17:35:44.476097] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:36.832 Malloc3 00:14:36.832 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc3 -n 2 00:14:36.832 [2024-10-17 17:35:44.689615] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:36.832 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:36.832 Asynchronous Event Request test 00:14:36.832 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:36.832 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:36.832 Registering asynchronous event callbacks... 00:14:36.832 Starting namespace attribute notice tests for all controllers... 00:14:36.832 /var/run/vfio-user/domain/vfio-user1/1: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:14:36.832 aer_cb - Changed Namespace 00:14:36.832 Cleaning up... 00:14:37.093 [ 00:14:37.093 { 00:14:37.093 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:37.093 "subtype": "Discovery", 00:14:37.093 "listen_addresses": [], 00:14:37.093 "allow_any_host": true, 00:14:37.093 "hosts": [] 00:14:37.093 }, 00:14:37.093 { 00:14:37.093 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:37.093 "subtype": "NVMe", 00:14:37.093 "listen_addresses": [ 00:14:37.093 { 00:14:37.093 "trtype": "VFIOUSER", 00:14:37.093 "adrfam": "IPv4", 00:14:37.093 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:37.093 "trsvcid": "0" 00:14:37.093 } 00:14:37.093 ], 00:14:37.093 "allow_any_host": true, 00:14:37.093 "hosts": [], 00:14:37.093 "serial_number": "SPDK1", 00:14:37.093 "model_number": "SPDK bdev Controller", 00:14:37.093 "max_namespaces": 32, 00:14:37.093 "min_cntlid": 1, 00:14:37.093 "max_cntlid": 65519, 00:14:37.093 "namespaces": [ 00:14:37.093 { 00:14:37.093 "nsid": 1, 00:14:37.093 "bdev_name": "Malloc1", 00:14:37.093 "name": "Malloc1", 00:14:37.093 "nguid": "98E788A861DA4B87A0F94BD5380D6C9A", 00:14:37.093 "uuid": "98e788a8-61da-4b87-a0f9-4bd5380d6c9a" 00:14:37.093 }, 00:14:37.093 { 00:14:37.093 "nsid": 2, 00:14:37.093 "bdev_name": "Malloc3", 00:14:37.093 "name": "Malloc3", 00:14:37.093 "nguid": "003088F788C944E4B6D9027982BC54D8", 00:14:37.093 "uuid": "003088f7-88c9-44e4-b6d9-027982bc54d8" 00:14:37.093 } 00:14:37.093 ] 00:14:37.093 }, 00:14:37.093 { 00:14:37.093 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:37.093 "subtype": "NVMe", 00:14:37.093 "listen_addresses": [ 00:14:37.093 { 00:14:37.093 "trtype": "VFIOUSER", 00:14:37.093 "adrfam": "IPv4", 00:14:37.094 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:37.094 "trsvcid": "0" 00:14:37.094 } 00:14:37.094 ], 00:14:37.094 "allow_any_host": true, 00:14:37.094 "hosts": [], 00:14:37.094 "serial_number": "SPDK2", 00:14:37.094 "model_number": "SPDK bdev Controller", 00:14:37.094 "max_namespaces": 32, 00:14:37.094 "min_cntlid": 1, 00:14:37.094 "max_cntlid": 65519, 00:14:37.094 "namespaces": [ 00:14:37.094 { 00:14:37.094 "nsid": 1, 00:14:37.094 "bdev_name": "Malloc2", 00:14:37.094 "name": "Malloc2", 00:14:37.094 "nguid": "6FA7B43C200F40C5BE19519108931ED7", 00:14:37.094 "uuid": "6fa7b43c-200f-40c5-be19-519108931ed7" 00:14:37.094 } 00:14:37.094 ] 00:14:37.094 } 00:14:37.094 ] 00:14:37.094 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 2114 00:14:37.094 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:14:37.094 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user2/2 00:14:37.094 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode2 00:14:37.094 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -L nvme -L nvme_vfio -L vfio_pci 00:14:37.094 [2024-10-17 17:35:44.921321] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:14:37.094 [2024-10-17 17:35:44.921363] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2176 ] 00:14:37.094 [2024-10-17 17:35:44.947338] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user2/2 00:14:37.094 [2024-10-17 17:35:44.955863] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:14:37.094 [2024-10-17 17:35:44.955882] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7f0657889000 00:14:37.094 [2024-10-17 17:35:44.956867] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.957872] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.958878] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.959886] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.960892] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.961901] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.962908] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.963917] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:37.094 [2024-10-17 17:35:44.964924] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:14:37.094 [2024-10-17 17:35:44.964933] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7f065787e000 00:14:37.094 [2024-10-17 17:35:44.965844] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:14:37.094 [2024-10-17 17:35:44.975223] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user2/2/cntrl Setup Successfully 00:14:37.094 [2024-10-17 17:35:44.975243] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to connect adminq (no timeout) 00:14:37.094 [2024-10-17 17:35:44.980303] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:14:37.094 [2024-10-17 17:35:44.980335] nvme_pcie_common.c: 134:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:14:37.094 [2024-10-17 17:35:44.980391] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for connect adminq (no timeout) 00:14:37.094 [2024-10-17 17:35:44.980404] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read vs (no timeout) 00:14:37.094 [2024-10-17 17:35:44.980407] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read vs wait for vs (no timeout) 00:14:37.094 [2024-10-17 17:35:44.981306] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x8, value 0x10300 00:14:37.094 [2024-10-17 17:35:44.981314] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read cap (no timeout) 00:14:37.094 [2024-10-17 17:35:44.981318] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read cap wait for cap (no timeout) 00:14:37.094 [2024-10-17 17:35:44.982311] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:14:37.094 [2024-10-17 17:35:44.982318] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to check en (no timeout) 00:14:37.094 [2024-10-17 17:35:44.982323] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to check en wait for cc (timeout 15000 ms) 00:14:37.094 [2024-10-17 17:35:44.983314] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x0 00:14:37.094 [2024-10-17 17:35:44.983321] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:14:37.094 [2024-10-17 17:35:44.984322] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x0 00:14:37.094 [2024-10-17 17:35:44.984329] nvme_ctrlr.c:3893:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CC.EN = 0 && CSTS.RDY = 0 00:14:37.094 [2024-10-17 17:35:44.984333] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to controller is disabled (timeout 15000 ms) 00:14:37.094 [2024-10-17 17:35:44.984337] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:14:37.094 [2024-10-17 17:35:44.984441] nvme_ctrlr.c:4091:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Setting CC.EN = 1 00:14:37.094 [2024-10-17 17:35:44.984447] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:14:37.094 [2024-10-17 17:35:44.984451] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x28, value 0x2000003c0000 00:14:37.094 [2024-10-17 17:35:44.985328] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x30, value 0x2000003be000 00:14:37.094 [2024-10-17 17:35:44.986337] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x24, value 0xff00ff 00:14:37.094 [2024-10-17 17:35:44.987345] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:14:37.094 [2024-10-17 17:35:44.988351] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:37.094 [2024-10-17 17:35:44.988384] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:14:37.094 [2024-10-17 17:35:44.989365] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x1 00:14:37.094 [2024-10-17 17:35:44.989372] nvme_ctrlr.c:3928:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:14:37.094 [2024-10-17 17:35:44.989375] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to reset admin queue (timeout 30000 ms) 00:14:37.094 [2024-10-17 17:35:44.989390] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify controller (no timeout) 00:14:37.094 [2024-10-17 17:35:44.989395] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify controller (timeout 30000 ms) 00:14:37.094 [2024-10-17 17:35:44.989404] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:37.094 [2024-10-17 17:35:44.989407] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:37.094 [2024-10-17 17:35:44.989410] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:37.094 [2024-10-17 17:35:44.989418] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:37.094 [2024-10-17 17:35:44.996696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:14:37.094 [2024-10-17 17:35:44.996705] nvme_ctrlr.c:2077:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] transport max_xfer_size 131072 00:14:37.094 [2024-10-17 17:35:44.996709] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] MDTS max_xfer_size 131072 00:14:37.094 [2024-10-17 17:35:44.996712] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CNTLID 0x0001 00:14:37.094 [2024-10-17 17:35:44.996715] nvme_ctrlr.c:2095:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:14:37.094 [2024-10-17 17:35:44.996718] nvme_ctrlr.c:2108:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] transport max_sges 1 00:14:37.094 [2024-10-17 17:35:44.996721] nvme_ctrlr.c:2123:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] fuses compare and write: 1 00:14:37.094 [2024-10-17 17:35:44.996724] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to configure AER (timeout 30000 ms) 00:14:37.094 [2024-10-17 17:35:44.996730] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for configure aer (timeout 30000 ms) 00:14:37.094 [2024-10-17 17:35:44.996738] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:14:37.094 [2024-10-17 17:35:45.004695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:14:37.094 [2024-10-17 17:35:45.004705] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:14:37.094 [2024-10-17 17:35:45.004712] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:14:37.094 [2024-10-17 17:35:45.004718] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:14:37.094 [2024-10-17 17:35:45.004724] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:14:37.094 [2024-10-17 17:35:45.004727] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set keep alive timeout (timeout 30000 ms) 00:14:37.095 [2024-10-17 17:35:45.004734] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:14:37.095 [2024-10-17 17:35:45.004741] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.012696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.012703] nvme_ctrlr.c:3034:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Controller adjusted keep alive timeout to 0 ms 00:14:37.356 [2024-10-17 17:35:45.012706] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify controller iocs specific (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.012711] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set number of queues (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.012717] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for set number of queues (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.012724] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.020697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.020743] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify active ns (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.020749] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify active ns (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.020754] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:14:37.356 [2024-10-17 17:35:45.020758] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:14:37.356 [2024-10-17 17:35:45.020760] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:37.356 [2024-10-17 17:35:45.020765] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.028695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.028704] nvme_ctrlr.c:4722:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Namespace 1 was added 00:14:37.356 [2024-10-17 17:35:45.028710] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify ns (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.028715] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify ns (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.028722] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:37.356 [2024-10-17 17:35:45.028725] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:37.356 [2024-10-17 17:35:45.028728] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:37.356 [2024-10-17 17:35:45.028732] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.036695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.036706] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify namespace id descriptors (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.036712] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.036717] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:37.356 [2024-10-17 17:35:45.036721] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:37.356 [2024-10-17 17:35:45.036723] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:37.356 [2024-10-17 17:35:45.036728] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.044695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.044703] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify ns iocs specific (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.044708] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set supported log pages (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.044714] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set supported features (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.044718] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set host behavior support feature (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.044722] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set doorbell buffer config (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.044726] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set host ID (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.044729] nvme_ctrlr.c:3134:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] NVMe-oF transport - not sending Set Features - Host ID 00:14:37.356 [2024-10-17 17:35:45.044733] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to transport ready (timeout 30000 ms) 00:14:37.356 [2024-10-17 17:35:45.044736] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to ready (no timeout) 00:14:37.356 [2024-10-17 17:35:45.044750] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.052695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.052706] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.060697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.060707] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.068695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.068706] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:14:37.356 [2024-10-17 17:35:45.076695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:14:37.356 [2024-10-17 17:35:45.076708] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:14:37.357 [2024-10-17 17:35:45.076711] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:14:37.357 [2024-10-17 17:35:45.076714] nvme_pcie_common.c:1241:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:14:37.357 [2024-10-17 17:35:45.076716] nvme_pcie_common.c:1257:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:14:37.357 [2024-10-17 17:35:45.076719] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:14:37.357 [2024-10-17 17:35:45.076723] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:14:37.357 [2024-10-17 17:35:45.076728] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:14:37.357 [2024-10-17 17:35:45.076731] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:14:37.357 [2024-10-17 17:35:45.076734] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:37.357 [2024-10-17 17:35:45.076738] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:14:37.357 [2024-10-17 17:35:45.076743] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:14:37.357 [2024-10-17 17:35:45.076746] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:37.357 [2024-10-17 17:35:45.076749] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:37.357 [2024-10-17 17:35:45.076753] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:37.357 [2024-10-17 17:35:45.076758] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:14:37.357 [2024-10-17 17:35:45.076761] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:14:37.357 [2024-10-17 17:35:45.076764] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:14:37.357 [2024-10-17 17:35:45.076768] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:14:37.357 [2024-10-17 17:35:45.084698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:14:37.357 [2024-10-17 17:35:45.084709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:14:37.357 [2024-10-17 17:35:45.084717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:14:37.357 [2024-10-17 17:35:45.084722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:14:37.357 ===================================================== 00:14:37.357 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:37.357 ===================================================== 00:14:37.357 Controller Capabilities/Features 00:14:37.357 ================================ 00:14:37.357 Vendor ID: 4e58 00:14:37.357 Subsystem Vendor ID: 4e58 00:14:37.357 Serial Number: SPDK2 00:14:37.357 Model Number: SPDK bdev Controller 00:14:37.357 Firmware Version: 25.01 00:14:37.357 Recommended Arb Burst: 6 00:14:37.357 IEEE OUI Identifier: 8d 6b 50 00:14:37.357 Multi-path I/O 00:14:37.357 May have multiple subsystem ports: Yes 00:14:37.357 May have multiple controllers: Yes 00:14:37.357 Associated with SR-IOV VF: No 00:14:37.357 Max Data Transfer Size: 131072 00:14:37.357 Max Number of Namespaces: 32 00:14:37.357 Max Number of I/O Queues: 127 00:14:37.357 NVMe Specification Version (VS): 1.3 00:14:37.357 NVMe Specification Version (Identify): 1.3 00:14:37.357 Maximum Queue Entries: 256 00:14:37.357 Contiguous Queues Required: Yes 00:14:37.357 Arbitration Mechanisms Supported 00:14:37.357 Weighted Round Robin: Not Supported 00:14:37.357 Vendor Specific: Not Supported 00:14:37.357 Reset Timeout: 15000 ms 00:14:37.357 Doorbell Stride: 4 bytes 00:14:37.357 NVM Subsystem Reset: Not Supported 00:14:37.357 Command Sets Supported 00:14:37.357 NVM Command Set: Supported 00:14:37.357 Boot Partition: Not Supported 00:14:37.357 Memory Page Size Minimum: 4096 bytes 00:14:37.357 Memory Page Size Maximum: 4096 bytes 00:14:37.357 Persistent Memory Region: Not Supported 00:14:37.357 Optional Asynchronous Events Supported 00:14:37.357 Namespace Attribute Notices: Supported 00:14:37.357 Firmware Activation Notices: Not Supported 00:14:37.357 ANA Change Notices: Not Supported 00:14:37.357 PLE Aggregate Log Change Notices: Not Supported 00:14:37.357 LBA Status Info Alert Notices: Not Supported 00:14:37.357 EGE Aggregate Log Change Notices: Not Supported 00:14:37.357 Normal NVM Subsystem Shutdown event: Not Supported 00:14:37.357 Zone Descriptor Change Notices: Not Supported 00:14:37.357 Discovery Log Change Notices: Not Supported 00:14:37.357 Controller Attributes 00:14:37.357 128-bit Host Identifier: Supported 00:14:37.357 Non-Operational Permissive Mode: Not Supported 00:14:37.357 NVM Sets: Not Supported 00:14:37.357 Read Recovery Levels: Not Supported 00:14:37.357 Endurance Groups: Not Supported 00:14:37.357 Predictable Latency Mode: Not Supported 00:14:37.357 Traffic Based Keep ALive: Not Supported 00:14:37.357 Namespace Granularity: Not Supported 00:14:37.357 SQ Associations: Not Supported 00:14:37.357 UUID List: Not Supported 00:14:37.357 Multi-Domain Subsystem: Not Supported 00:14:37.357 Fixed Capacity Management: Not Supported 00:14:37.357 Variable Capacity Management: Not Supported 00:14:37.357 Delete Endurance Group: Not Supported 00:14:37.357 Delete NVM Set: Not Supported 00:14:37.357 Extended LBA Formats Supported: Not Supported 00:14:37.357 Flexible Data Placement Supported: Not Supported 00:14:37.357 00:14:37.357 Controller Memory Buffer Support 00:14:37.357 ================================ 00:14:37.357 Supported: No 00:14:37.357 00:14:37.357 Persistent Memory Region Support 00:14:37.357 ================================ 00:14:37.357 Supported: No 00:14:37.357 00:14:37.357 Admin Command Set Attributes 00:14:37.357 ============================ 00:14:37.357 Security Send/Receive: Not Supported 00:14:37.357 Format NVM: Not Supported 00:14:37.357 Firmware Activate/Download: Not Supported 00:14:37.357 Namespace Management: Not Supported 00:14:37.357 Device Self-Test: Not Supported 00:14:37.357 Directives: Not Supported 00:14:37.357 NVMe-MI: Not Supported 00:14:37.357 Virtualization Management: Not Supported 00:14:37.357 Doorbell Buffer Config: Not Supported 00:14:37.357 Get LBA Status Capability: Not Supported 00:14:37.357 Command & Feature Lockdown Capability: Not Supported 00:14:37.357 Abort Command Limit: 4 00:14:37.357 Async Event Request Limit: 4 00:14:37.357 Number of Firmware Slots: N/A 00:14:37.357 Firmware Slot 1 Read-Only: N/A 00:14:37.357 Firmware Activation Without Reset: N/A 00:14:37.357 Multiple Update Detection Support: N/A 00:14:37.357 Firmware Update Granularity: No Information Provided 00:14:37.357 Per-Namespace SMART Log: No 00:14:37.357 Asymmetric Namespace Access Log Page: Not Supported 00:14:37.357 Subsystem NQN: nqn.2019-07.io.spdk:cnode2 00:14:37.357 Command Effects Log Page: Supported 00:14:37.357 Get Log Page Extended Data: Supported 00:14:37.357 Telemetry Log Pages: Not Supported 00:14:37.357 Persistent Event Log Pages: Not Supported 00:14:37.357 Supported Log Pages Log Page: May Support 00:14:37.357 Commands Supported & Effects Log Page: Not Supported 00:14:37.357 Feature Identifiers & Effects Log Page:May Support 00:14:37.357 NVMe-MI Commands & Effects Log Page: May Support 00:14:37.357 Data Area 4 for Telemetry Log: Not Supported 00:14:37.357 Error Log Page Entries Supported: 128 00:14:37.357 Keep Alive: Supported 00:14:37.357 Keep Alive Granularity: 10000 ms 00:14:37.357 00:14:37.357 NVM Command Set Attributes 00:14:37.357 ========================== 00:14:37.357 Submission Queue Entry Size 00:14:37.357 Max: 64 00:14:37.357 Min: 64 00:14:37.357 Completion Queue Entry Size 00:14:37.357 Max: 16 00:14:37.357 Min: 16 00:14:37.357 Number of Namespaces: 32 00:14:37.357 Compare Command: Supported 00:14:37.357 Write Uncorrectable Command: Not Supported 00:14:37.357 Dataset Management Command: Supported 00:14:37.357 Write Zeroes Command: Supported 00:14:37.357 Set Features Save Field: Not Supported 00:14:37.357 Reservations: Not Supported 00:14:37.357 Timestamp: Not Supported 00:14:37.357 Copy: Supported 00:14:37.357 Volatile Write Cache: Present 00:14:37.357 Atomic Write Unit (Normal): 1 00:14:37.357 Atomic Write Unit (PFail): 1 00:14:37.357 Atomic Compare & Write Unit: 1 00:14:37.357 Fused Compare & Write: Supported 00:14:37.357 Scatter-Gather List 00:14:37.357 SGL Command Set: Supported (Dword aligned) 00:14:37.357 SGL Keyed: Not Supported 00:14:37.357 SGL Bit Bucket Descriptor: Not Supported 00:14:37.357 SGL Metadata Pointer: Not Supported 00:14:37.357 Oversized SGL: Not Supported 00:14:37.358 SGL Metadata Address: Not Supported 00:14:37.358 SGL Offset: Not Supported 00:14:37.358 Transport SGL Data Block: Not Supported 00:14:37.358 Replay Protected Memory Block: Not Supported 00:14:37.358 00:14:37.358 Firmware Slot Information 00:14:37.358 ========================= 00:14:37.358 Active slot: 1 00:14:37.358 Slot 1 Firmware Revision: 25.01 00:14:37.358 00:14:37.358 00:14:37.358 Commands Supported and Effects 00:14:37.358 ============================== 00:14:37.358 Admin Commands 00:14:37.358 -------------- 00:14:37.358 Get Log Page (02h): Supported 00:14:37.358 Identify (06h): Supported 00:14:37.358 Abort (08h): Supported 00:14:37.358 Set Features (09h): Supported 00:14:37.358 Get Features (0Ah): Supported 00:14:37.358 Asynchronous Event Request (0Ch): Supported 00:14:37.358 Keep Alive (18h): Supported 00:14:37.358 I/O Commands 00:14:37.358 ------------ 00:14:37.358 Flush (00h): Supported LBA-Change 00:14:37.358 Write (01h): Supported LBA-Change 00:14:37.358 Read (02h): Supported 00:14:37.358 Compare (05h): Supported 00:14:37.358 Write Zeroes (08h): Supported LBA-Change 00:14:37.358 Dataset Management (09h): Supported LBA-Change 00:14:37.358 Copy (19h): Supported LBA-Change 00:14:37.358 00:14:37.358 Error Log 00:14:37.358 ========= 00:14:37.358 00:14:37.358 Arbitration 00:14:37.358 =========== 00:14:37.358 Arbitration Burst: 1 00:14:37.358 00:14:37.358 Power Management 00:14:37.358 ================ 00:14:37.358 Number of Power States: 1 00:14:37.358 Current Power State: Power State #0 00:14:37.358 Power State #0: 00:14:37.358 Max Power: 0.00 W 00:14:37.358 Non-Operational State: Operational 00:14:37.358 Entry Latency: Not Reported 00:14:37.358 Exit Latency: Not Reported 00:14:37.358 Relative Read Throughput: 0 00:14:37.358 Relative Read Latency: 0 00:14:37.358 Relative Write Throughput: 0 00:14:37.358 Relative Write Latency: 0 00:14:37.358 Idle Power: Not Reported 00:14:37.358 Active Power: Not Reported 00:14:37.358 Non-Operational Permissive Mode: Not Supported 00:14:37.358 00:14:37.358 Health Information 00:14:37.358 ================== 00:14:37.358 Critical Warnings: 00:14:37.358 Available Spare Space: OK 00:14:37.358 Temperature: OK 00:14:37.358 Device Reliability: OK 00:14:37.358 Read Only: No 00:14:37.358 Volatile Memory Backup: OK 00:14:37.358 Current Temperature: 0 Kelvin (-273 Celsius) 00:14:37.358 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:14:37.358 Available Spare: 0% 00:14:37.358 Available Sp[2024-10-17 17:35:45.084793] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:14:37.358 [2024-10-17 17:35:45.092695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:14:37.358 [2024-10-17 17:35:45.092719] nvme_ctrlr.c:4386:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Prepare to destruct SSD 00:14:37.358 [2024-10-17 17:35:45.092727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:37.358 [2024-10-17 17:35:45.092732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:37.358 [2024-10-17 17:35:45.092736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:37.358 [2024-10-17 17:35:45.092741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:37.358 [2024-10-17 17:35:45.092778] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:14:37.358 [2024-10-17 17:35:45.092786] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x464001 00:14:37.358 [2024-10-17 17:35:45.093782] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:37.358 [2024-10-17 17:35:45.093818] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] RTD3E = 0 us 00:14:37.358 [2024-10-17 17:35:45.093823] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] shutdown timeout = 10000 ms 00:14:37.358 [2024-10-17 17:35:45.094783] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x9 00:14:37.358 [2024-10-17 17:35:45.094793] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] shutdown complete in 0 milliseconds 00:14:37.358 [2024-10-17 17:35:45.094840] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user2/2/cntrl 00:14:37.358 [2024-10-17 17:35:45.095803] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:14:37.358 are Threshold: 0% 00:14:37.358 Life Percentage Used: 0% 00:14:37.358 Data Units Read: 0 00:14:37.358 Data Units Written: 0 00:14:37.358 Host Read Commands: 0 00:14:37.358 Host Write Commands: 0 00:14:37.358 Controller Busy Time: 0 minutes 00:14:37.358 Power Cycles: 0 00:14:37.358 Power On Hours: 0 hours 00:14:37.358 Unsafe Shutdowns: 0 00:14:37.358 Unrecoverable Media Errors: 0 00:14:37.358 Lifetime Error Log Entries: 0 00:14:37.358 Warning Temperature Time: 0 minutes 00:14:37.358 Critical Temperature Time: 0 minutes 00:14:37.358 00:14:37.358 Number of Queues 00:14:37.358 ================ 00:14:37.358 Number of I/O Submission Queues: 127 00:14:37.358 Number of I/O Completion Queues: 127 00:14:37.358 00:14:37.358 Active Namespaces 00:14:37.358 ================= 00:14:37.358 Namespace ID:1 00:14:37.358 Error Recovery Timeout: Unlimited 00:14:37.358 Command Set Identifier: NVM (00h) 00:14:37.358 Deallocate: Supported 00:14:37.358 Deallocated/Unwritten Error: Not Supported 00:14:37.358 Deallocated Read Value: Unknown 00:14:37.358 Deallocate in Write Zeroes: Not Supported 00:14:37.358 Deallocated Guard Field: 0xFFFF 00:14:37.358 Flush: Supported 00:14:37.358 Reservation: Supported 00:14:37.358 Namespace Sharing Capabilities: Multiple Controllers 00:14:37.358 Size (in LBAs): 131072 (0GiB) 00:14:37.358 Capacity (in LBAs): 131072 (0GiB) 00:14:37.358 Utilization (in LBAs): 131072 (0GiB) 00:14:37.358 NGUID: 6FA7B43C200F40C5BE19519108931ED7 00:14:37.358 UUID: 6fa7b43c-200f-40c5-be19-519108931ed7 00:14:37.358 Thin Provisioning: Not Supported 00:14:37.358 Per-NS Atomic Units: Yes 00:14:37.358 Atomic Boundary Size (Normal): 0 00:14:37.358 Atomic Boundary Size (PFail): 0 00:14:37.358 Atomic Boundary Offset: 0 00:14:37.358 Maximum Single Source Range Length: 65535 00:14:37.358 Maximum Copy Length: 65535 00:14:37.358 Maximum Source Range Count: 1 00:14:37.358 NGUID/EUI64 Never Reused: No 00:14:37.358 Namespace Write Protected: No 00:14:37.358 Number of LBA Formats: 1 00:14:37.358 Current LBA Format: LBA Format #00 00:14:37.358 LBA Format #00: Data Size: 512 Metadata Size: 0 00:14:37.358 00:14:37.358 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:14:37.618 [2024-10-17 17:35:45.276721] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:42.896 Initializing NVMe Controllers 00:14:42.896 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:42.896 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:14:42.896 Initialization complete. Launching workers. 00:14:42.896 ======================================================== 00:14:42.896 Latency(us) 00:14:42.896 Device Information : IOPS MiB/s Average min max 00:14:42.896 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 39977.07 156.16 3201.51 845.78 8752.84 00:14:42.896 ======================================================== 00:14:42.896 Total : 39977.07 156.16 3201.51 845.78 8752.84 00:14:42.896 00:14:42.896 [2024-10-17 17:35:50.381891] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:42.896 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:14:42.896 [2024-10-17 17:35:50.560440] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:48.177 Initializing NVMe Controllers 00:14:48.177 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:48.177 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:14:48.177 Initialization complete. Launching workers. 00:14:48.177 ======================================================== 00:14:48.177 Latency(us) 00:14:48.177 Device Information : IOPS MiB/s Average min max 00:14:48.177 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 40063.80 156.50 3194.85 849.25 8990.63 00:14:48.177 ======================================================== 00:14:48.177 Total : 40063.80 156.50 3194.85 849.25 8990.63 00:14:48.177 00:14:48.177 [2024-10-17 17:35:55.580017] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:48.177 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:14:48.177 [2024-10-17 17:35:55.764163] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:53.455 [2024-10-17 17:36:00.903776] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:53.455 Initializing NVMe Controllers 00:14:53.455 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:53.455 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:53.455 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 1 00:14:53.455 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 2 00:14:53.455 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 3 00:14:53.455 Initialization complete. Launching workers. 00:14:53.455 Starting thread on core 2 00:14:53.455 Starting thread on core 3 00:14:53.455 Starting thread on core 1 00:14:53.455 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -d 256 -g 00:14:53.455 [2024-10-17 17:36:01.142106] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:56.754 [2024-10-17 17:36:04.204820] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:56.754 Initializing NVMe Controllers 00:14:56.754 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:56.754 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:56.754 Associating SPDK bdev Controller (SPDK2 ) with lcore 0 00:14:56.754 Associating SPDK bdev Controller (SPDK2 ) with lcore 1 00:14:56.754 Associating SPDK bdev Controller (SPDK2 ) with lcore 2 00:14:56.754 Associating SPDK bdev Controller (SPDK2 ) with lcore 3 00:14:56.754 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:14:56.754 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:14:56.754 Initialization complete. Launching workers. 00:14:56.754 Starting thread on core 1 with urgent priority queue 00:14:56.754 Starting thread on core 2 with urgent priority queue 00:14:56.754 Starting thread on core 3 with urgent priority queue 00:14:56.754 Starting thread on core 0 with urgent priority queue 00:14:56.754 SPDK bdev Controller (SPDK2 ) core 0: 11788.00 IO/s 8.48 secs/100000 ios 00:14:56.754 SPDK bdev Controller (SPDK2 ) core 1: 6637.33 IO/s 15.07 secs/100000 ios 00:14:56.754 SPDK bdev Controller (SPDK2 ) core 2: 14715.00 IO/s 6.80 secs/100000 ios 00:14:56.754 SPDK bdev Controller (SPDK2 ) core 3: 8797.33 IO/s 11.37 secs/100000 ios 00:14:56.754 ======================================================== 00:14:56.754 00:14:56.754 17:36:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:14:56.754 [2024-10-17 17:36:04.440737] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:56.754 Initializing NVMe Controllers 00:14:56.754 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:56.754 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:56.754 Namespace ID: 1 size: 0GB 00:14:56.754 Initialization complete. 00:14:56.754 INFO: using host memory buffer for IO 00:14:56.754 Hello world! 00:14:56.754 [2024-10-17 17:36:04.450804] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:56.754 17:36:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:14:57.014 [2024-10-17 17:36:04.675370] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:57.954 Initializing NVMe Controllers 00:14:57.954 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:57.954 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:57.954 Initialization complete. Launching workers. 00:14:57.954 submit (in ns) avg, min, max = 7547.8, 2839.2, 3998060.0 00:14:57.954 complete (in ns) avg, min, max = 15757.2, 1638.3, 7986141.7 00:14:57.954 00:14:57.954 Submit histogram 00:14:57.954 ================ 00:14:57.954 Range in us Cumulative Count 00:14:57.954 2.827 - 2.840: 0.0049% ( 1) 00:14:57.954 2.840 - 2.853: 0.0692% ( 13) 00:14:57.954 2.853 - 2.867: 0.4847% ( 84) 00:14:57.954 2.867 - 2.880: 2.1220% ( 331) 00:14:57.954 2.880 - 2.893: 5.5597% ( 695) 00:14:57.954 2.893 - 2.907: 9.6800% ( 833) 00:14:57.954 2.907 - 2.920: 14.7697% ( 1029) 00:14:57.954 2.920 - 2.933: 19.9733% ( 1052) 00:14:57.954 2.933 - 2.947: 25.1274% ( 1042) 00:14:57.954 2.947 - 2.960: 30.1776% ( 1021) 00:14:57.954 2.960 - 2.973: 35.3317% ( 1042) 00:14:57.954 2.973 - 2.987: 40.9012% ( 1126) 00:14:57.954 2.987 - 3.000: 47.0297% ( 1239) 00:14:57.954 3.000 - 3.013: 53.1830% ( 1244) 00:14:57.954 3.013 - 3.027: 60.7162% ( 1523) 00:14:57.954 3.027 - 3.040: 68.6501% ( 1604) 00:14:57.954 3.040 - 3.053: 76.6929% ( 1626) 00:14:57.954 3.053 - 3.067: 84.1371% ( 1505) 00:14:57.954 3.067 - 3.080: 90.5080% ( 1288) 00:14:57.954 3.080 - 3.093: 94.8014% ( 868) 00:14:57.954 3.093 - 3.107: 97.3784% ( 521) 00:14:57.954 3.107 - 3.120: 98.7980% ( 287) 00:14:57.954 3.120 - 3.133: 99.3867% ( 119) 00:14:57.954 3.133 - 3.147: 99.5647% ( 36) 00:14:57.954 3.147 - 3.160: 99.5993% ( 7) 00:14:57.954 3.160 - 3.173: 99.6142% ( 3) 00:14:57.954 3.173 - 3.187: 99.6241% ( 2) 00:14:57.954 3.200 - 3.213: 99.6290% ( 1) 00:14:57.954 3.627 - 3.653: 99.6340% ( 1) 00:14:57.954 3.733 - 3.760: 99.6389% ( 1) 00:14:57.954 3.867 - 3.893: 99.6439% ( 1) 00:14:57.954 4.267 - 4.293: 99.6538% ( 2) 00:14:57.954 4.667 - 4.693: 99.6587% ( 1) 00:14:57.954 4.693 - 4.720: 99.6636% ( 1) 00:14:57.954 4.720 - 4.747: 99.6686% ( 1) 00:14:57.954 4.747 - 4.773: 99.6785% ( 2) 00:14:57.954 4.773 - 4.800: 99.6834% ( 1) 00:14:57.954 4.800 - 4.827: 99.6884% ( 1) 00:14:57.954 4.827 - 4.853: 99.6983% ( 2) 00:14:57.954 4.853 - 4.880: 99.7032% ( 1) 00:14:57.954 4.880 - 4.907: 99.7082% ( 1) 00:14:57.954 4.907 - 4.933: 99.7131% ( 1) 00:14:57.954 4.960 - 4.987: 99.7230% ( 2) 00:14:57.954 4.987 - 5.013: 99.7280% ( 1) 00:14:57.954 5.013 - 5.040: 99.7329% ( 1) 00:14:57.954 5.067 - 5.093: 99.7428% ( 2) 00:14:57.954 5.093 - 5.120: 99.7477% ( 1) 00:14:57.954 5.147 - 5.173: 99.7576% ( 2) 00:14:57.954 5.173 - 5.200: 99.7626% ( 1) 00:14:57.954 5.200 - 5.227: 99.7675% ( 1) 00:14:57.954 5.227 - 5.253: 99.7725% ( 1) 00:14:57.954 5.253 - 5.280: 99.7824% ( 2) 00:14:57.954 5.280 - 5.307: 99.7873% ( 1) 00:14:57.954 5.307 - 5.333: 99.8021% ( 3) 00:14:57.954 5.360 - 5.387: 99.8120% ( 2) 00:14:57.954 5.573 - 5.600: 99.8170% ( 1) 00:14:57.954 5.600 - 5.627: 99.8219% ( 1) 00:14:57.954 5.627 - 5.653: 99.8269% ( 1) 00:14:57.954 5.653 - 5.680: 99.8368% ( 2) 00:14:57.954 5.733 - 5.760: 99.8417% ( 1) 00:14:57.954 5.760 - 5.787: 99.8467% ( 1) 00:14:57.954 5.867 - 5.893: 99.8516% ( 1) 00:14:57.954 5.973 - 6.000: 99.8566% ( 1) 00:14:57.954 6.027 - 6.053: 99.8615% ( 1) 00:14:57.954 6.160 - 6.187: 99.8664% ( 1) 00:14:57.954 6.187 - 6.213: 99.8714% ( 1) 00:14:57.954 6.213 - 6.240: 99.8763% ( 1) 00:14:57.954 6.987 - 7.040: 99.8813% ( 1) 00:14:57.954 7.787 - 7.840: 99.8862% ( 1) 00:14:57.954 3986.773 - 4014.080: 100.0000% ( 23) 00:14:57.954 00:14:57.954 Complete histogram 00:14:57.954 ================== 00:14:57.954 Range in us Cumulative Count 00:14:57.954 1.633 - 1.640: 0.0049% ( 1) 00:14:57.954 1.640 - 1.647: 0.2523% ( 50) 00:14:57.954 1.647 - 1.653: 0.8607% ( 123) 00:14:57.954 1.653 - 1.660: 0.9398% ( 16) 00:14:57.954 1.660 - 1.667: 1.1228% ( 37) 00:14:57.954 1.667 - 1.673: 1.2415% ( 24) 00:14:57.954 1.673 - 1.680: 1.3009% ( 12) 00:14:57.954 1.680 - [2024-10-17 17:36:05.770268] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:57.954 1.687: 1.3058% ( 1) 00:14:57.954 1.687 - 1.693: 1.3207% ( 3) 00:14:57.954 1.693 - 1.700: 2.2803% ( 194) 00:14:57.954 1.700 - 1.707: 29.6582% ( 5535) 00:14:57.954 1.707 - 1.720: 48.6670% ( 3843) 00:14:57.954 1.720 - 1.733: 74.7143% ( 5266) 00:14:57.954 1.733 - 1.747: 82.6928% ( 1613) 00:14:57.954 1.747 - 1.760: 84.6268% ( 391) 00:14:57.954 1.760 - 1.773: 87.7479% ( 631) 00:14:57.954 1.773 - 1.787: 92.2986% ( 920) 00:14:57.954 1.787 - 1.800: 96.2210% ( 793) 00:14:57.954 1.800 - 1.813: 98.4716% ( 455) 00:14:57.954 1.813 - 1.827: 99.2432% ( 156) 00:14:57.954 1.827 - 1.840: 99.3916% ( 30) 00:14:57.954 1.840 - 1.853: 99.4361% ( 9) 00:14:57.954 1.853 - 1.867: 99.4460% ( 2) 00:14:57.954 1.867 - 1.880: 99.4510% ( 1) 00:14:57.954 1.973 - 1.987: 99.4559% ( 1) 00:14:57.954 1.987 - 2.000: 99.4608% ( 1) 00:14:57.954 2.013 - 2.027: 99.4658% ( 1) 00:14:57.954 2.080 - 2.093: 99.4707% ( 1) 00:14:57.954 2.147 - 2.160: 99.4757% ( 1) 00:14:57.954 2.187 - 2.200: 99.4806% ( 1) 00:14:57.954 3.493 - 3.520: 99.4856% ( 1) 00:14:57.954 3.520 - 3.547: 99.4905% ( 1) 00:14:57.954 3.573 - 3.600: 99.4955% ( 1) 00:14:57.954 3.760 - 3.787: 99.5004% ( 1) 00:14:57.954 3.813 - 3.840: 99.5054% ( 1) 00:14:57.954 3.840 - 3.867: 99.5202% ( 3) 00:14:57.954 3.867 - 3.893: 99.5301% ( 2) 00:14:57.954 3.920 - 3.947: 99.5400% ( 2) 00:14:57.954 3.973 - 4.000: 99.5449% ( 1) 00:14:57.954 4.000 - 4.027: 99.5499% ( 1) 00:14:57.954 4.053 - 4.080: 99.5548% ( 1) 00:14:57.954 4.107 - 4.133: 99.5598% ( 1) 00:14:57.954 4.133 - 4.160: 99.5647% ( 1) 00:14:57.954 4.187 - 4.213: 99.5697% ( 1) 00:14:57.954 4.213 - 4.240: 99.5845% ( 3) 00:14:57.954 4.240 - 4.267: 99.5895% ( 1) 00:14:57.954 4.320 - 4.347: 99.5944% ( 1) 00:14:57.955 4.507 - 4.533: 99.5993% ( 1) 00:14:57.955 4.640 - 4.667: 99.6043% ( 1) 00:14:57.955 4.693 - 4.720: 99.6092% ( 1) 00:14:57.955 4.720 - 4.747: 99.6142% ( 1) 00:14:57.955 4.800 - 4.827: 99.6241% ( 2) 00:14:57.955 4.987 - 5.013: 99.6290% ( 1) 00:14:57.955 5.013 - 5.040: 99.6340% ( 1) 00:14:57.955 5.067 - 5.093: 99.6389% ( 1) 00:14:57.955 9.333 - 9.387: 99.6439% ( 1) 00:14:57.955 10.347 - 10.400: 99.6488% ( 1) 00:14:57.955 11.360 - 11.413: 99.6538% ( 1) 00:14:57.955 3986.773 - 4014.080: 99.9951% ( 69) 00:14:57.955 7973.547 - 8028.160: 100.0000% ( 1) 00:14:57.955 00:14:57.955 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user2/2 nqn.2019-07.io.spdk:cnode2 2 00:14:57.955 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user2/2 00:14:57.955 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode2 00:14:57.955 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc4 00:14:57.955 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:58.215 [ 00:14:58.216 { 00:14:58.216 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:58.216 "subtype": "Discovery", 00:14:58.216 "listen_addresses": [], 00:14:58.216 "allow_any_host": true, 00:14:58.216 "hosts": [] 00:14:58.216 }, 00:14:58.216 { 00:14:58.216 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:58.216 "subtype": "NVMe", 00:14:58.216 "listen_addresses": [ 00:14:58.216 { 00:14:58.216 "trtype": "VFIOUSER", 00:14:58.216 "adrfam": "IPv4", 00:14:58.216 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:58.216 "trsvcid": "0" 00:14:58.216 } 00:14:58.216 ], 00:14:58.216 "allow_any_host": true, 00:14:58.216 "hosts": [], 00:14:58.216 "serial_number": "SPDK1", 00:14:58.216 "model_number": "SPDK bdev Controller", 00:14:58.216 "max_namespaces": 32, 00:14:58.216 "min_cntlid": 1, 00:14:58.216 "max_cntlid": 65519, 00:14:58.216 "namespaces": [ 00:14:58.216 { 00:14:58.216 "nsid": 1, 00:14:58.216 "bdev_name": "Malloc1", 00:14:58.216 "name": "Malloc1", 00:14:58.216 "nguid": "98E788A861DA4B87A0F94BD5380D6C9A", 00:14:58.216 "uuid": "98e788a8-61da-4b87-a0f9-4bd5380d6c9a" 00:14:58.216 }, 00:14:58.216 { 00:14:58.216 "nsid": 2, 00:14:58.216 "bdev_name": "Malloc3", 00:14:58.216 "name": "Malloc3", 00:14:58.216 "nguid": "003088F788C944E4B6D9027982BC54D8", 00:14:58.216 "uuid": "003088f7-88c9-44e4-b6d9-027982bc54d8" 00:14:58.216 } 00:14:58.216 ] 00:14:58.216 }, 00:14:58.216 { 00:14:58.216 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:58.216 "subtype": "NVMe", 00:14:58.216 "listen_addresses": [ 00:14:58.216 { 00:14:58.216 "trtype": "VFIOUSER", 00:14:58.216 "adrfam": "IPv4", 00:14:58.216 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:58.216 "trsvcid": "0" 00:14:58.216 } 00:14:58.216 ], 00:14:58.216 "allow_any_host": true, 00:14:58.216 "hosts": [], 00:14:58.216 "serial_number": "SPDK2", 00:14:58.216 "model_number": "SPDK bdev Controller", 00:14:58.216 "max_namespaces": 32, 00:14:58.216 "min_cntlid": 1, 00:14:58.216 "max_cntlid": 65519, 00:14:58.216 "namespaces": [ 00:14:58.216 { 00:14:58.216 "nsid": 1, 00:14:58.216 "bdev_name": "Malloc2", 00:14:58.216 "name": "Malloc2", 00:14:58.216 "nguid": "6FA7B43C200F40C5BE19519108931ED7", 00:14:58.216 "uuid": "6fa7b43c-200f-40c5-be19-519108931ed7" 00:14:58.216 } 00:14:58.216 ] 00:14:58.216 } 00:14:58.216 ] 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=6638 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1265 -- # local i=0 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -n 2 -g -t /tmp/aer_touch_file 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # return 0 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:14:58.216 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc4 00:14:58.477 [2024-10-17 17:36:06.139050] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:58.477 Malloc4 00:14:58.477 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc4 -n 2 00:14:58.477 [2024-10-17 17:36:06.343492] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:58.477 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:58.477 Asynchronous Event Request test 00:14:58.477 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:58.477 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:58.477 Registering asynchronous event callbacks... 00:14:58.477 Starting namespace attribute notice tests for all controllers... 00:14:58.477 /var/run/vfio-user/domain/vfio-user2/2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:14:58.477 aer_cb - Changed Namespace 00:14:58.477 Cleaning up... 00:14:58.737 [ 00:14:58.737 { 00:14:58.737 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:58.737 "subtype": "Discovery", 00:14:58.737 "listen_addresses": [], 00:14:58.737 "allow_any_host": true, 00:14:58.737 "hosts": [] 00:14:58.737 }, 00:14:58.737 { 00:14:58.737 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:58.737 "subtype": "NVMe", 00:14:58.737 "listen_addresses": [ 00:14:58.737 { 00:14:58.737 "trtype": "VFIOUSER", 00:14:58.737 "adrfam": "IPv4", 00:14:58.737 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:58.737 "trsvcid": "0" 00:14:58.737 } 00:14:58.737 ], 00:14:58.737 "allow_any_host": true, 00:14:58.737 "hosts": [], 00:14:58.737 "serial_number": "SPDK1", 00:14:58.737 "model_number": "SPDK bdev Controller", 00:14:58.737 "max_namespaces": 32, 00:14:58.737 "min_cntlid": 1, 00:14:58.737 "max_cntlid": 65519, 00:14:58.737 "namespaces": [ 00:14:58.737 { 00:14:58.737 "nsid": 1, 00:14:58.737 "bdev_name": "Malloc1", 00:14:58.737 "name": "Malloc1", 00:14:58.737 "nguid": "98E788A861DA4B87A0F94BD5380D6C9A", 00:14:58.737 "uuid": "98e788a8-61da-4b87-a0f9-4bd5380d6c9a" 00:14:58.737 }, 00:14:58.737 { 00:14:58.737 "nsid": 2, 00:14:58.737 "bdev_name": "Malloc3", 00:14:58.737 "name": "Malloc3", 00:14:58.737 "nguid": "003088F788C944E4B6D9027982BC54D8", 00:14:58.737 "uuid": "003088f7-88c9-44e4-b6d9-027982bc54d8" 00:14:58.737 } 00:14:58.737 ] 00:14:58.737 }, 00:14:58.737 { 00:14:58.737 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:58.737 "subtype": "NVMe", 00:14:58.737 "listen_addresses": [ 00:14:58.737 { 00:14:58.737 "trtype": "VFIOUSER", 00:14:58.737 "adrfam": "IPv4", 00:14:58.737 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:58.737 "trsvcid": "0" 00:14:58.737 } 00:14:58.737 ], 00:14:58.737 "allow_any_host": true, 00:14:58.737 "hosts": [], 00:14:58.738 "serial_number": "SPDK2", 00:14:58.738 "model_number": "SPDK bdev Controller", 00:14:58.738 "max_namespaces": 32, 00:14:58.738 "min_cntlid": 1, 00:14:58.738 "max_cntlid": 65519, 00:14:58.738 "namespaces": [ 00:14:58.738 { 00:14:58.738 "nsid": 1, 00:14:58.738 "bdev_name": "Malloc2", 00:14:58.738 "name": "Malloc2", 00:14:58.738 "nguid": "6FA7B43C200F40C5BE19519108931ED7", 00:14:58.738 "uuid": "6fa7b43c-200f-40c5-be19-519108931ed7" 00:14:58.738 }, 00:14:58.738 { 00:14:58.738 "nsid": 2, 00:14:58.738 "bdev_name": "Malloc4", 00:14:58.738 "name": "Malloc4", 00:14:58.738 "nguid": "D8C839BF098748ECBC8B007C9591E627", 00:14:58.738 "uuid": "d8c839bf-0987-48ec-bc8b-007c9591e627" 00:14:58.738 } 00:14:58.738 ] 00:14:58.738 } 00:14:58.738 ] 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 6638 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@105 -- # stop_nvmf_vfio_user 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 4190770 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@950 -- # '[' -z 4190770 ']' 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # kill -0 4190770 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # uname 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 4190770 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@968 -- # echo 'killing process with pid 4190770' 00:14:58.738 killing process with pid 4190770 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@969 -- # kill 4190770 00:14:58.738 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@974 -- # wait 4190770 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@108 -- # setup_nvmf_vfio_user --interrupt-mode '-M -I' 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args=--interrupt-mode 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local 'transport_args=-M -I' 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=6807 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 6807' 00:14:58.999 Process pid: 6807 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' --interrupt-mode 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 6807 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@831 -- # '[' -z 6807 ']' 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:58.999 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:58.999 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:14:58.999 [2024-10-17 17:36:06.825801] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:14:58.999 [2024-10-17 17:36:06.826735] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:14:58.999 [2024-10-17 17:36:06.826780] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:58.999 [2024-10-17 17:36:06.903921] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:59.259 [2024-10-17 17:36:06.932964] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:59.259 [2024-10-17 17:36:06.932998] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:59.259 [2024-10-17 17:36:06.933004] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:59.259 [2024-10-17 17:36:06.933008] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:59.259 [2024-10-17 17:36:06.933013] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:59.259 [2024-10-17 17:36:06.934425] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:59.259 [2024-10-17 17:36:06.934579] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:59.259 [2024-10-17 17:36:06.934694] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:59.259 [2024-10-17 17:36:06.934712] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:59.259 [2024-10-17 17:36:06.985469] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:14:59.259 [2024-10-17 17:36:06.986440] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:14:59.259 [2024-10-17 17:36:06.986764] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:14:59.259 [2024-10-17 17:36:06.987912] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:14:59.259 [2024-10-17 17:36:06.987951] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:14:59.841 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:59.841 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # return 0 00:14:59.841 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:15:00.781 17:36:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER -M -I 00:15:01.042 17:36:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:15:01.042 17:36:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:15:01.042 17:36:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:15:01.042 17:36:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:15:01.042 17:36:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:15:01.302 Malloc1 00:15:01.302 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:15:01.563 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:15:01.563 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:15:01.824 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:15:01.824 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:15:01.824 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:15:02.085 Malloc2 00:15:02.085 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:15:02.085 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:15:02.350 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:15:02.609 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@109 -- # stop_nvmf_vfio_user 00:15:02.609 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 6807 00:15:02.609 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@950 -- # '[' -z 6807 ']' 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # kill -0 6807 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # uname 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 6807 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@968 -- # echo 'killing process with pid 6807' 00:15:02.610 killing process with pid 6807 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@969 -- # kill 6807 00:15:02.610 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@974 -- # wait 6807 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:15:02.870 00:15:02.870 real 0m50.825s 00:15:02.870 user 3m14.725s 00:15:02.870 sys 0m2.771s 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:15:02.870 ************************************ 00:15:02.870 END TEST nvmf_vfio_user 00:15:02.870 ************************************ 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@32 -- # run_test nvmf_vfio_user_nvme_compliance /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:02.870 ************************************ 00:15:02.870 START TEST nvmf_vfio_user_nvme_compliance 00:15:02.870 ************************************ 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:15:02.870 * Looking for test storage... 00:15:02.870 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1691 -- # lcov --version 00:15:02.870 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # IFS=.-: 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # read -ra ver1 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # IFS=.-: 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # read -ra ver2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@338 -- # local 'op=<' 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@340 -- # ver1_l=2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@341 -- # ver2_l=1 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@344 -- # case "$op" in 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@345 -- # : 1 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # decimal 1 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=1 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 1 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # ver1[v]=1 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # decimal 2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # ver2[v]=2 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # return 0 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:15:03.132 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:03.132 --rc genhtml_branch_coverage=1 00:15:03.132 --rc genhtml_function_coverage=1 00:15:03.132 --rc genhtml_legend=1 00:15:03.132 --rc geninfo_all_blocks=1 00:15:03.132 --rc geninfo_unexecuted_blocks=1 00:15:03.132 00:15:03.132 ' 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:15:03.132 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:03.132 --rc genhtml_branch_coverage=1 00:15:03.132 --rc genhtml_function_coverage=1 00:15:03.132 --rc genhtml_legend=1 00:15:03.132 --rc geninfo_all_blocks=1 00:15:03.132 --rc geninfo_unexecuted_blocks=1 00:15:03.132 00:15:03.132 ' 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:15:03.132 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:03.132 --rc genhtml_branch_coverage=1 00:15:03.132 --rc genhtml_function_coverage=1 00:15:03.132 --rc genhtml_legend=1 00:15:03.132 --rc geninfo_all_blocks=1 00:15:03.132 --rc geninfo_unexecuted_blocks=1 00:15:03.132 00:15:03.132 ' 00:15:03.132 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:15:03.132 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:03.132 --rc genhtml_branch_coverage=1 00:15:03.132 --rc genhtml_function_coverage=1 00:15:03.132 --rc genhtml_legend=1 00:15:03.132 --rc geninfo_all_blocks=1 00:15:03.133 --rc geninfo_unexecuted_blocks=1 00:15:03.133 00:15:03.133 ' 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # uname -s 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@15 -- # shopt -s extglob 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@5 -- # export PATH 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@51 -- # : 0 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:03.133 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@11 -- # MALLOC_BDEV_SIZE=64 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # export TEST_TRANSPORT=VFIOUSER 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # TEST_TRANSPORT=VFIOUSER 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@16 -- # rm -rf /var/run/vfio-user 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@20 -- # nvmfpid=7948 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@21 -- # echo 'Process pid: 7948' 00:15:03.133 Process pid: 7948 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@23 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@24 -- # waitforlisten 7948 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@831 -- # '[' -z 7948 ']' 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:03.133 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:03.133 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:15:03.133 [2024-10-17 17:36:10.923910] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:15:03.133 [2024-10-17 17:36:10.923992] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:03.133 [2024-10-17 17:36:11.007131] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:15:03.133 [2024-10-17 17:36:11.049861] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:03.133 [2024-10-17 17:36:11.049903] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:03.133 [2024-10-17 17:36:11.049909] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:03.133 [2024-10-17 17:36:11.049914] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:03.133 [2024-10-17 17:36:11.049919] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:03.394 [2024-10-17 17:36:11.051574] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:03.394 [2024-10-17 17:36:11.051750] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:03.394 [2024-10-17 17:36:11.051752] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:03.964 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:03.964 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@864 -- # return 0 00:15:03.964 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@26 -- # sleep 1 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@28 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@29 -- # traddr=/var/run/vfio-user 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@31 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@33 -- # mkdir -p /var/run/vfio-user 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@35 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:15:04.907 malloc0 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@36 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk -m 32 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@37 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@38 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.907 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:15:05.167 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.167 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/nvme_compliance -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user subnqn:nqn.2021-09.io.spdk:cnode0' 00:15:05.167 00:15:05.167 00:15:05.167 CUnit - A unit testing framework for C - Version 2.1-3 00:15:05.167 http://cunit.sourceforge.net/ 00:15:05.167 00:15:05.167 00:15:05.167 Suite: nvme_compliance 00:15:05.167 Test: admin_identify_ctrlr_verify_dptr ...[2024-10-17 17:36:12.972187] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.167 [2024-10-17 17:36:12.973471] vfio_user.c: 804:nvme_cmd_map_prps: *ERROR*: no PRP2, 3072 remaining 00:15:05.167 [2024-10-17 17:36:12.973483] vfio_user.c:5507:map_admin_cmd_req: *ERROR*: /var/run/vfio-user: map Admin Opc 6 failed 00:15:05.167 [2024-10-17 17:36:12.973488] vfio_user.c:5600:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x6 failed 00:15:05.167 [2024-10-17 17:36:12.975207] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.167 passed 00:15:05.167 Test: admin_identify_ctrlr_verify_fused ...[2024-10-17 17:36:13.052701] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.167 [2024-10-17 17:36:13.055717] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.167 passed 00:15:05.427 Test: admin_identify_ns ...[2024-10-17 17:36:13.134258] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.427 [2024-10-17 17:36:13.193704] ctrlr.c:2750:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 0 00:15:05.427 [2024-10-17 17:36:13.201700] ctrlr.c:2750:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 4294967295 00:15:05.427 [2024-10-17 17:36:13.222784] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.427 passed 00:15:05.427 Test: admin_get_features_mandatory_features ...[2024-10-17 17:36:13.297014] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.427 [2024-10-17 17:36:13.300035] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.427 passed 00:15:05.687 Test: admin_get_features_optional_features ...[2024-10-17 17:36:13.379493] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.687 [2024-10-17 17:36:13.382520] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.687 passed 00:15:05.687 Test: admin_set_features_number_of_queues ...[2024-10-17 17:36:13.457265] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.687 [2024-10-17 17:36:13.561779] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.687 passed 00:15:05.947 Test: admin_get_log_page_mandatory_logs ...[2024-10-17 17:36:13.634994] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.947 [2024-10-17 17:36:13.638014] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.947 passed 00:15:05.947 Test: admin_get_log_page_with_lpo ...[2024-10-17 17:36:13.714757] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:05.947 [2024-10-17 17:36:13.782699] ctrlr.c:2697:nvmf_ctrlr_get_log_page: *ERROR*: Get log page: offset (516) > len (512) 00:15:05.947 [2024-10-17 17:36:13.795748] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:05.947 passed 00:15:06.207 Test: fabric_property_get ...[2024-10-17 17:36:13.870808] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.207 [2024-10-17 17:36:13.872017] vfio_user.c:5600:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x7f failed 00:15:06.207 [2024-10-17 17:36:13.873827] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.207 passed 00:15:06.207 Test: admin_delete_io_sq_use_admin_qid ...[2024-10-17 17:36:13.952303] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.207 [2024-10-17 17:36:13.953510] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:0 does not exist 00:15:06.207 [2024-10-17 17:36:13.955328] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.207 passed 00:15:06.207 Test: admin_delete_io_sq_delete_sq_twice ...[2024-10-17 17:36:14.030072] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.207 [2024-10-17 17:36:14.113698] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:15:06.466 [2024-10-17 17:36:14.129698] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:15:06.466 [2024-10-17 17:36:14.134771] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.466 passed 00:15:06.466 Test: admin_delete_io_cq_use_admin_qid ...[2024-10-17 17:36:14.207987] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.466 [2024-10-17 17:36:14.209180] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O cqid:0 does not exist 00:15:06.466 [2024-10-17 17:36:14.211002] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.466 passed 00:15:06.466 Test: admin_delete_io_cq_delete_cq_first ...[2024-10-17 17:36:14.287754] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.466 [2024-10-17 17:36:14.363701] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:15:06.726 [2024-10-17 17:36:14.387698] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:15:06.726 [2024-10-17 17:36:14.392766] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.726 passed 00:15:06.726 Test: admin_create_io_cq_verify_iv_pc ...[2024-10-17 17:36:14.467942] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.726 [2024-10-17 17:36:14.469142] vfio_user.c:2158:handle_create_io_cq: *ERROR*: /var/run/vfio-user: IV is too big 00:15:06.726 [2024-10-17 17:36:14.469158] vfio_user.c:2152:handle_create_io_cq: *ERROR*: /var/run/vfio-user: non-PC CQ not supported 00:15:06.726 [2024-10-17 17:36:14.470959] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.726 passed 00:15:06.726 Test: admin_create_io_sq_verify_qsize_cqid ...[2024-10-17 17:36:14.545672] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.726 [2024-10-17 17:36:14.634696] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 1 00:15:06.726 [2024-10-17 17:36:14.642696] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 257 00:15:06.986 [2024-10-17 17:36:14.650695] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:0 00:15:06.987 [2024-10-17 17:36:14.658699] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:128 00:15:06.987 [2024-10-17 17:36:14.687769] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.987 passed 00:15:06.987 Test: admin_create_io_sq_verify_pc ...[2024-10-17 17:36:14.763790] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:06.987 [2024-10-17 17:36:14.781702] vfio_user.c:2051:handle_create_io_sq: *ERROR*: /var/run/vfio-user: non-PC SQ not supported 00:15:06.987 [2024-10-17 17:36:14.798939] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:06.987 passed 00:15:06.987 Test: admin_create_io_qp_max_qps ...[2024-10-17 17:36:14.873391] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:08.367 [2024-10-17 17:36:15.969698] nvme_ctrlr.c:5504:spdk_nvme_ctrlr_alloc_qid: *ERROR*: [/var/run/vfio-user] No free I/O queue IDs 00:15:08.627 [2024-10-17 17:36:16.350547] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:08.627 passed 00:15:08.627 Test: admin_create_io_sq_shared_cq ...[2024-10-17 17:36:16.427406] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:15:08.888 [2024-10-17 17:36:16.559702] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:15:08.888 [2024-10-17 17:36:16.596747] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:15:08.888 passed 00:15:08.888 00:15:08.888 Run Summary: Type Total Ran Passed Failed Inactive 00:15:08.888 suites 1 1 n/a 0 0 00:15:08.888 tests 18 18 18 0 0 00:15:08.888 asserts 360 360 360 0 n/a 00:15:08.888 00:15:08.888 Elapsed time = 1.486 seconds 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@42 -- # killprocess 7948 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@950 -- # '[' -z 7948 ']' 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@954 -- # kill -0 7948 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@955 -- # uname 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 7948 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@968 -- # echo 'killing process with pid 7948' 00:15:08.888 killing process with pid 7948 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@969 -- # kill 7948 00:15:08.888 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@974 -- # wait 7948 00:15:09.149 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@44 -- # rm -rf /var/run/vfio-user 00:15:09.149 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:15:09.149 00:15:09.149 real 0m6.188s 00:15:09.149 user 0m17.517s 00:15:09.149 sys 0m0.547s 00:15:09.149 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:09.149 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:15:09.150 ************************************ 00:15:09.150 END TEST nvmf_vfio_user_nvme_compliance 00:15:09.150 ************************************ 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@33 -- # run_test nvmf_vfio_user_fuzz /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:09.150 ************************************ 00:15:09.150 START TEST nvmf_vfio_user_fuzz 00:15:09.150 ************************************ 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:15:09.150 * Looking for test storage... 00:15:09.150 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1691 -- # lcov --version 00:15:09.150 17:36:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # IFS=.-: 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # read -ra ver1 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # IFS=.-: 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # read -ra ver2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@338 -- # local 'op=<' 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@340 -- # ver1_l=2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@341 -- # ver2_l=1 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@344 -- # case "$op" in 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@345 -- # : 1 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # decimal 1 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=1 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 1 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # ver1[v]=1 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # decimal 2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # ver2[v]=2 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:09.410 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # return 0 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:15:09.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:09.411 --rc genhtml_branch_coverage=1 00:15:09.411 --rc genhtml_function_coverage=1 00:15:09.411 --rc genhtml_legend=1 00:15:09.411 --rc geninfo_all_blocks=1 00:15:09.411 --rc geninfo_unexecuted_blocks=1 00:15:09.411 00:15:09.411 ' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:15:09.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:09.411 --rc genhtml_branch_coverage=1 00:15:09.411 --rc genhtml_function_coverage=1 00:15:09.411 --rc genhtml_legend=1 00:15:09.411 --rc geninfo_all_blocks=1 00:15:09.411 --rc geninfo_unexecuted_blocks=1 00:15:09.411 00:15:09.411 ' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:15:09.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:09.411 --rc genhtml_branch_coverage=1 00:15:09.411 --rc genhtml_function_coverage=1 00:15:09.411 --rc genhtml_legend=1 00:15:09.411 --rc geninfo_all_blocks=1 00:15:09.411 --rc geninfo_unexecuted_blocks=1 00:15:09.411 00:15:09.411 ' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:15:09.411 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:09.411 --rc genhtml_branch_coverage=1 00:15:09.411 --rc genhtml_function_coverage=1 00:15:09.411 --rc genhtml_legend=1 00:15:09.411 --rc geninfo_all_blocks=1 00:15:09.411 --rc geninfo_unexecuted_blocks=1 00:15:09.411 00:15:09.411 ' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # uname -s 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@15 -- # shopt -s extglob 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@5 -- # export PATH 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@51 -- # : 0 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:09.411 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@12 -- # MALLOC_BDEV_SIZE=64 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@15 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@16 -- # traddr=/var/run/vfio-user 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@20 -- # rm -rf /var/run/vfio-user 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@24 -- # nvmfpid=9428 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@25 -- # echo 'Process pid: 9428' 00:15:09.411 Process pid: 9428 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@27 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@28 -- # waitforlisten 9428 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@831 -- # '[' -z 9428 ']' 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:09.411 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:09.411 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:10.353 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:10.353 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@864 -- # return 0 00:15:10.353 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@30 -- # sleep 1 00:15:11.294 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@32 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:15:11.294 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.294 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:11.294 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.294 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@34 -- # mkdir -p /var/run/vfio-user 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:11.294 malloc0 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@39 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.294 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:11.295 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.295 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@41 -- # trid='trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' 00:15:11.295 17:36:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -t 30 -S 123456 -F 'trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' -N -a 00:15:43.419 Fuzzing completed. Shutting down the fuzz application 00:15:43.419 00:15:43.419 Dumping successful admin opcodes: 00:15:43.419 8, 9, 10, 24, 00:15:43.419 Dumping successful io opcodes: 00:15:43.419 0, 00:15:43.419 NS: 0x20000081ef00 I/O qp, Total commands completed: 1303673, total successful commands: 5104, random_seed: 2562774656 00:15:43.419 NS: 0x20000081ef00 admin qp, Total commands completed: 311239, total successful commands: 2507, random_seed: 800193216 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@44 -- # rpc_cmd nvmf_delete_subsystem nqn.2021-09.io.spdk:cnode0 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@46 -- # killprocess 9428 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@950 -- # '[' -z 9428 ']' 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@954 -- # kill -0 9428 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@955 -- # uname 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 9428 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@968 -- # echo 'killing process with pid 9428' 00:15:43.419 killing process with pid 9428 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@969 -- # kill 9428 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@974 -- # wait 9428 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@48 -- # rm -rf /var/run/vfio-user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_log.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_tgt_output.txt 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@50 -- # trap - SIGINT SIGTERM EXIT 00:15:43.419 00:15:43.419 real 0m32.792s 00:15:43.419 user 0m35.545s 00:15:43.419 sys 0m26.742s 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:15:43.419 ************************************ 00:15:43.419 END TEST nvmf_vfio_user_fuzz 00:15:43.419 ************************************ 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@37 -- # run_test nvmf_auth_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:43.419 ************************************ 00:15:43.419 START TEST nvmf_auth_target 00:15:43.419 ************************************ 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:15:43.419 * Looking for test storage... 00:15:43.419 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1691 -- # lcov --version 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # IFS=.-: 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # read -ra ver1 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # IFS=.-: 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # read -ra ver2 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@338 -- # local 'op=<' 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@340 -- # ver1_l=2 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@341 -- # ver2_l=1 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@344 -- # case "$op" in 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@345 -- # : 1 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # decimal 1 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=1 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 1 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # ver1[v]=1 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # decimal 2 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=2 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:43.419 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 2 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # ver2[v]=2 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # return 0 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:15:43.420 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:43.420 --rc genhtml_branch_coverage=1 00:15:43.420 --rc genhtml_function_coverage=1 00:15:43.420 --rc genhtml_legend=1 00:15:43.420 --rc geninfo_all_blocks=1 00:15:43.420 --rc geninfo_unexecuted_blocks=1 00:15:43.420 00:15:43.420 ' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:15:43.420 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:43.420 --rc genhtml_branch_coverage=1 00:15:43.420 --rc genhtml_function_coverage=1 00:15:43.420 --rc genhtml_legend=1 00:15:43.420 --rc geninfo_all_blocks=1 00:15:43.420 --rc geninfo_unexecuted_blocks=1 00:15:43.420 00:15:43.420 ' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:15:43.420 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:43.420 --rc genhtml_branch_coverage=1 00:15:43.420 --rc genhtml_function_coverage=1 00:15:43.420 --rc genhtml_legend=1 00:15:43.420 --rc geninfo_all_blocks=1 00:15:43.420 --rc geninfo_unexecuted_blocks=1 00:15:43.420 00:15:43.420 ' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:15:43.420 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:43.420 --rc genhtml_branch_coverage=1 00:15:43.420 --rc genhtml_function_coverage=1 00:15:43.420 --rc genhtml_legend=1 00:15:43.420 --rc geninfo_all_blocks=1 00:15:43.420 --rc geninfo_unexecuted_blocks=1 00:15:43.420 00:15:43.420 ' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # uname -s 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@15 -- # shopt -s extglob 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@5 -- # export PATH 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@51 -- # : 0 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:43.420 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@14 -- # dhgroups=("null" "ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@15 -- # subnqn=nqn.2024-03.io.spdk:cnode0 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@16 -- # hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@17 -- # hostsock=/var/tmp/host.sock 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # keys=() 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # ckeys=() 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@86 -- # nvmftestinit 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:43.420 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:43.420 17:36:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:15:43.420 17:36:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:15:43.420 17:36:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@309 -- # xtrace_disable 00:15:43.420 17:36:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # pci_devs=() 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # net_devs=() 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # e810=() 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # local -ga e810 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # x722=() 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # local -ga x722 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # mlx=() 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # local -ga mlx 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:15:50.112 Found 0000:31:00.0 (0x8086 - 0x159b) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:15:50.112 Found 0000:31:00.1 (0x8086 - 0x159b) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:15:50.112 Found net devices under 0000:31:00.0: cvl_0_0 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:15:50.112 Found net devices under 0000:31:00.1: cvl_0_1 00:15:50.112 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # is_hw=yes 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:50.113 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:50.113 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.625 ms 00:15:50.113 00:15:50.113 --- 10.0.0.2 ping statistics --- 00:15:50.113 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:50.113 rtt min/avg/max/mdev = 0.625/0.625/0.625/0.000 ms 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:50.113 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:50.113 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.276 ms 00:15:50.113 00:15:50.113 --- 10.0.0.1 ping statistics --- 00:15:50.113 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:50.113 rtt min/avg/max/mdev = 0.276/0.276/0.276/0.000 ms 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@448 -- # return 0 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@87 -- # nvmfappstart -L nvmf_auth 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # nvmfpid=19452 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # waitforlisten 19452 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvmf_auth 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 19452 ']' 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:50.113 17:36:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@89 -- # hostpid=19666 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@91 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 2 -r /var/tmp/host.sock -L nvme_auth 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key null 48 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=null 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=48 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=55be4be5e231ba9b461201b694c8982e3b1862847268e90a 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.iNU 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 55be4be5e231ba9b461201b694c8982e3b1862847268e90a 0 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 55be4be5e231ba9b461201b694c8982e3b1862847268e90a 0 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=55be4be5e231ba9b461201b694c8982e3b1862847268e90a 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=0 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.iNU 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.iNU 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # keys[0]=/tmp/spdk.key-null.iNU 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key sha512 64 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha512 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=64 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=6b1e5786cefe49614d2d0aa80f5693c67023ac14d2aca02d2675eab225d094c5 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.IWh 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 6b1e5786cefe49614d2d0aa80f5693c67023ac14d2aca02d2675eab225d094c5 3 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 6b1e5786cefe49614d2d0aa80f5693c67023ac14d2aca02d2675eab225d094c5 3 00:15:50.686 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=6b1e5786cefe49614d2d0aa80f5693c67023ac14d2aca02d2675eab225d094c5 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=3 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.IWh 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.IWh 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # ckeys[0]=/tmp/spdk.key-sha512.IWh 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha256 32 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha256 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=32 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=8b204b3cb97f8e820d8cb21ef4bbd758 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.EzL 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 8b204b3cb97f8e820d8cb21ef4bbd758 1 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 8b204b3cb97f8e820d8cb21ef4bbd758 1 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=8b204b3cb97f8e820d8cb21ef4bbd758 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=1 00:15:50.687 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.EzL 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.EzL 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # keys[1]=/tmp/spdk.key-sha256.EzL 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha384 48 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha384 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=48 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=4b13faf528e2ec69d68451dde799d7c561d3344d1c157d63 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.cFu 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 4b13faf528e2ec69d68451dde799d7c561d3344d1c157d63 2 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 4b13faf528e2ec69d68451dde799d7c561d3344d1c157d63 2 00:15:50.948 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=4b13faf528e2ec69d68451dde799d7c561d3344d1c157d63 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=2 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.cFu 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.cFu 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # ckeys[1]=/tmp/spdk.key-sha384.cFu 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha384 48 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha384 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=48 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=b2f0c14d96274914a9d235a56800a8cf4b872760cc065200 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.0Y8 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key b2f0c14d96274914a9d235a56800a8cf4b872760cc065200 2 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 b2f0c14d96274914a9d235a56800a8cf4b872760cc065200 2 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=b2f0c14d96274914a9d235a56800a8cf4b872760cc065200 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=2 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.0Y8 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.0Y8 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # keys[2]=/tmp/spdk.key-sha384.0Y8 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha256 32 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha256 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=32 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=1c7cfb6360f9886cb24d14e8ce696562 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.VIR 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 1c7cfb6360f9886cb24d14e8ce696562 1 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 1c7cfb6360f9886cb24d14e8ce696562 1 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=1c7cfb6360f9886cb24d14e8ce696562 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=1 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.VIR 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.VIR 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # ckeys[2]=/tmp/spdk.key-sha256.VIR 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # gen_dhchap_key sha512 64 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha512 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=64 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=eaa70a412ff208d056ee878defe5fa052939651759ac6b09abfc2b2147050aba 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.RrJ 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key eaa70a412ff208d056ee878defe5fa052939651759ac6b09abfc2b2147050aba 3 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 eaa70a412ff208d056ee878defe5fa052939651759ac6b09abfc2b2147050aba 3 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=eaa70a412ff208d056ee878defe5fa052939651759ac6b09abfc2b2147050aba 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=3 00:15:50.949 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.RrJ 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.RrJ 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # keys[3]=/tmp/spdk.key-sha512.RrJ 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # ckeys[3]= 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@99 -- # waitforlisten 19452 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 19452 ']' 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:51.210 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:51.210 17:36:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@100 -- # waitforlisten 19666 /var/tmp/host.sock 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 19666 ']' 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/host.sock 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:15:51.210 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:51.210 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@101 -- # rpc_cmd 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.iNU 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key0 /tmp/spdk.key-null.iNU 00:15:51.471 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key0 /tmp/spdk.key-null.iNU 00:15:51.732 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha512.IWh ]] 00:15:51.732 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.IWh 00:15:51.732 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.732 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:51.732 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.732 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey0 /tmp/spdk.key-sha512.IWh 00:15:51.732 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey0 /tmp/spdk.key-sha512.IWh 00:15:51.994 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:15:51.994 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.EzL 00:15:51.994 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.994 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:51.994 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.994 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key1 /tmp/spdk.key-sha256.EzL 00:15:51.994 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key1 /tmp/spdk.key-sha256.EzL 00:15:52.255 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha384.cFu ]] 00:15:52.255 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.cFu 00:15:52.255 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.255 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:52.255 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.255 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey1 /tmp/spdk.key-sha384.cFu 00:15:52.255 17:36:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey1 /tmp/spdk.key-sha384.cFu 00:15:52.255 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:15:52.255 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.0Y8 00:15:52.255 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.255 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:52.255 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.255 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key2 /tmp/spdk.key-sha384.0Y8 00:15:52.255 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key2 /tmp/spdk.key-sha384.0Y8 00:15:52.517 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha256.VIR ]] 00:15:52.517 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.VIR 00:15:52.517 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.517 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:52.517 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.517 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey2 /tmp/spdk.key-sha256.VIR 00:15:52.517 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey2 /tmp/spdk.key-sha256.VIR 00:15:52.778 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:15:52.778 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.RrJ 00:15:52.778 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.778 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:52.778 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.778 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key3 /tmp/spdk.key-sha512.RrJ 00:15:52.778 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key3 /tmp/spdk.key-sha512.RrJ 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n '' ]] 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 0 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:53.039 17:37:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:53.300 00:15:53.300 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:15:53.300 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:15:53.300 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:15:53.560 { 00:15:53.560 "cntlid": 1, 00:15:53.560 "qid": 0, 00:15:53.560 "state": "enabled", 00:15:53.560 "thread": "nvmf_tgt_poll_group_000", 00:15:53.560 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:15:53.560 "listen_address": { 00:15:53.560 "trtype": "TCP", 00:15:53.560 "adrfam": "IPv4", 00:15:53.560 "traddr": "10.0.0.2", 00:15:53.560 "trsvcid": "4420" 00:15:53.560 }, 00:15:53.560 "peer_address": { 00:15:53.560 "trtype": "TCP", 00:15:53.560 "adrfam": "IPv4", 00:15:53.560 "traddr": "10.0.0.1", 00:15:53.560 "trsvcid": "55150" 00:15:53.560 }, 00:15:53.560 "auth": { 00:15:53.560 "state": "completed", 00:15:53.560 "digest": "sha256", 00:15:53.560 "dhgroup": "null" 00:15:53.560 } 00:15:53.560 } 00:15:53.560 ]' 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:15:53.560 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:15:53.820 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:15:53.820 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:15:53.820 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:15:53.820 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:15:53.820 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:15:53.821 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:15:54.081 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:15:54.081 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:15:54.650 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:54.650 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 1 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.910 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:15:54.911 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:15:54.911 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:15:55.172 00:15:55.172 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:15:55.172 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:15:55.172 17:37:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:15:55.172 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:15:55.172 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:15:55.172 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.172 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:55.172 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.172 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:15:55.172 { 00:15:55.172 "cntlid": 3, 00:15:55.172 "qid": 0, 00:15:55.172 "state": "enabled", 00:15:55.172 "thread": "nvmf_tgt_poll_group_000", 00:15:55.172 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:15:55.172 "listen_address": { 00:15:55.172 "trtype": "TCP", 00:15:55.172 "adrfam": "IPv4", 00:15:55.172 "traddr": "10.0.0.2", 00:15:55.172 "trsvcid": "4420" 00:15:55.172 }, 00:15:55.172 "peer_address": { 00:15:55.172 "trtype": "TCP", 00:15:55.172 "adrfam": "IPv4", 00:15:55.172 "traddr": "10.0.0.1", 00:15:55.172 "trsvcid": "40594" 00:15:55.172 }, 00:15:55.172 "auth": { 00:15:55.172 "state": "completed", 00:15:55.172 "digest": "sha256", 00:15:55.172 "dhgroup": "null" 00:15:55.172 } 00:15:55.172 } 00:15:55.172 ]' 00:15:55.172 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:15:55.433 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:15:55.433 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:15:55.433 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:15:55.433 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:15:55.433 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:15:55.433 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:15:55.433 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:15:55.693 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:15:55.693 17:37:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:15:56.264 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:56.264 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 2 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:15:56.524 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:15:56.784 00:15:56.784 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:15:56.784 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:15:56.784 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:15:56.784 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:15:56.784 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:15:56.784 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.784 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:15:57.044 { 00:15:57.044 "cntlid": 5, 00:15:57.044 "qid": 0, 00:15:57.044 "state": "enabled", 00:15:57.044 "thread": "nvmf_tgt_poll_group_000", 00:15:57.044 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:15:57.044 "listen_address": { 00:15:57.044 "trtype": "TCP", 00:15:57.044 "adrfam": "IPv4", 00:15:57.044 "traddr": "10.0.0.2", 00:15:57.044 "trsvcid": "4420" 00:15:57.044 }, 00:15:57.044 "peer_address": { 00:15:57.044 "trtype": "TCP", 00:15:57.044 "adrfam": "IPv4", 00:15:57.044 "traddr": "10.0.0.1", 00:15:57.044 "trsvcid": "40638" 00:15:57.044 }, 00:15:57.044 "auth": { 00:15:57.044 "state": "completed", 00:15:57.044 "digest": "sha256", 00:15:57.044 "dhgroup": "null" 00:15:57.044 } 00:15:57.044 } 00:15:57.044 ]' 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:15:57.044 17:37:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:15:57.304 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:15:57.304 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:15:57.874 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:57.874 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 3 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:15:58.134 17:37:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:15:58.394 00:15:58.394 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:15:58.394 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:15:58.394 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:15:58.394 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:15:58.394 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:15:58.394 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.394 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:15:58.655 { 00:15:58.655 "cntlid": 7, 00:15:58.655 "qid": 0, 00:15:58.655 "state": "enabled", 00:15:58.655 "thread": "nvmf_tgt_poll_group_000", 00:15:58.655 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:15:58.655 "listen_address": { 00:15:58.655 "trtype": "TCP", 00:15:58.655 "adrfam": "IPv4", 00:15:58.655 "traddr": "10.0.0.2", 00:15:58.655 "trsvcid": "4420" 00:15:58.655 }, 00:15:58.655 "peer_address": { 00:15:58.655 "trtype": "TCP", 00:15:58.655 "adrfam": "IPv4", 00:15:58.655 "traddr": "10.0.0.1", 00:15:58.655 "trsvcid": "40664" 00:15:58.655 }, 00:15:58.655 "auth": { 00:15:58.655 "state": "completed", 00:15:58.655 "digest": "sha256", 00:15:58.655 "dhgroup": "null" 00:15:58.655 } 00:15:58.655 } 00:15:58.655 ]' 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:15:58.655 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:15:58.914 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:15:58.914 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:15:59.494 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:15:59.494 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 0 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:59.757 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:00.018 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:00.018 { 00:16:00.018 "cntlid": 9, 00:16:00.018 "qid": 0, 00:16:00.018 "state": "enabled", 00:16:00.018 "thread": "nvmf_tgt_poll_group_000", 00:16:00.018 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:00.018 "listen_address": { 00:16:00.018 "trtype": "TCP", 00:16:00.018 "adrfam": "IPv4", 00:16:00.018 "traddr": "10.0.0.2", 00:16:00.018 "trsvcid": "4420" 00:16:00.018 }, 00:16:00.018 "peer_address": { 00:16:00.018 "trtype": "TCP", 00:16:00.018 "adrfam": "IPv4", 00:16:00.018 "traddr": "10.0.0.1", 00:16:00.018 "trsvcid": "40690" 00:16:00.018 }, 00:16:00.018 "auth": { 00:16:00.018 "state": "completed", 00:16:00.018 "digest": "sha256", 00:16:00.018 "dhgroup": "ffdhe2048" 00:16:00.018 } 00:16:00.018 } 00:16:00.018 ]' 00:16:00.018 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:00.277 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:00.277 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:00.277 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:00.277 17:37:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:00.277 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:00.277 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:00.277 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:00.537 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:00.537 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:01.107 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:01.107 17:37:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 1 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:01.367 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:01.628 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:01.628 { 00:16:01.628 "cntlid": 11, 00:16:01.628 "qid": 0, 00:16:01.628 "state": "enabled", 00:16:01.628 "thread": "nvmf_tgt_poll_group_000", 00:16:01.628 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:01.628 "listen_address": { 00:16:01.628 "trtype": "TCP", 00:16:01.628 "adrfam": "IPv4", 00:16:01.628 "traddr": "10.0.0.2", 00:16:01.628 "trsvcid": "4420" 00:16:01.628 }, 00:16:01.628 "peer_address": { 00:16:01.628 "trtype": "TCP", 00:16:01.628 "adrfam": "IPv4", 00:16:01.628 "traddr": "10.0.0.1", 00:16:01.628 "trsvcid": "40726" 00:16:01.628 }, 00:16:01.628 "auth": { 00:16:01.628 "state": "completed", 00:16:01.628 "digest": "sha256", 00:16:01.628 "dhgroup": "ffdhe2048" 00:16:01.628 } 00:16:01.628 } 00:16:01.628 ]' 00:16:01.628 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:01.889 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:01.889 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:01.889 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:01.889 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:01.889 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:01.889 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:01.889 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:02.150 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:02.150 17:37:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:02.722 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:02.722 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 2 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:02.982 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:02.982 00:16:03.242 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:03.242 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:03.242 17:37:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:03.242 { 00:16:03.242 "cntlid": 13, 00:16:03.242 "qid": 0, 00:16:03.242 "state": "enabled", 00:16:03.242 "thread": "nvmf_tgt_poll_group_000", 00:16:03.242 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:03.242 "listen_address": { 00:16:03.242 "trtype": "TCP", 00:16:03.242 "adrfam": "IPv4", 00:16:03.242 "traddr": "10.0.0.2", 00:16:03.242 "trsvcid": "4420" 00:16:03.242 }, 00:16:03.242 "peer_address": { 00:16:03.242 "trtype": "TCP", 00:16:03.242 "adrfam": "IPv4", 00:16:03.242 "traddr": "10.0.0.1", 00:16:03.242 "trsvcid": "40736" 00:16:03.242 }, 00:16:03.242 "auth": { 00:16:03.242 "state": "completed", 00:16:03.242 "digest": "sha256", 00:16:03.242 "dhgroup": "ffdhe2048" 00:16:03.242 } 00:16:03.242 } 00:16:03.242 ]' 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:03.242 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:03.503 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:03.503 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:03.503 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:03.503 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:03.503 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:03.763 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:03.763 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:04.333 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:04.334 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:04.334 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:04.334 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.334 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:04.334 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.334 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:04.334 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:04.334 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 3 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:04.595 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:04.856 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:04.856 { 00:16:04.856 "cntlid": 15, 00:16:04.856 "qid": 0, 00:16:04.856 "state": "enabled", 00:16:04.856 "thread": "nvmf_tgt_poll_group_000", 00:16:04.856 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:04.856 "listen_address": { 00:16:04.856 "trtype": "TCP", 00:16:04.856 "adrfam": "IPv4", 00:16:04.856 "traddr": "10.0.0.2", 00:16:04.856 "trsvcid": "4420" 00:16:04.856 }, 00:16:04.856 "peer_address": { 00:16:04.856 "trtype": "TCP", 00:16:04.856 "adrfam": "IPv4", 00:16:04.856 "traddr": "10.0.0.1", 00:16:04.856 "trsvcid": "52058" 00:16:04.856 }, 00:16:04.856 "auth": { 00:16:04.856 "state": "completed", 00:16:04.856 "digest": "sha256", 00:16:04.856 "dhgroup": "ffdhe2048" 00:16:04.856 } 00:16:04.856 } 00:16:04.856 ]' 00:16:04.856 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:05.116 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:05.116 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:05.116 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:05.116 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:05.116 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:05.116 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:05.116 17:37:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:05.376 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:05.376 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:05.947 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:05.947 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 0 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:06.208 17:37:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:06.469 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.469 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:06.469 { 00:16:06.469 "cntlid": 17, 00:16:06.469 "qid": 0, 00:16:06.469 "state": "enabled", 00:16:06.469 "thread": "nvmf_tgt_poll_group_000", 00:16:06.469 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:06.469 "listen_address": { 00:16:06.469 "trtype": "TCP", 00:16:06.469 "adrfam": "IPv4", 00:16:06.469 "traddr": "10.0.0.2", 00:16:06.469 "trsvcid": "4420" 00:16:06.469 }, 00:16:06.469 "peer_address": { 00:16:06.469 "trtype": "TCP", 00:16:06.469 "adrfam": "IPv4", 00:16:06.469 "traddr": "10.0.0.1", 00:16:06.469 "trsvcid": "52082" 00:16:06.469 }, 00:16:06.469 "auth": { 00:16:06.469 "state": "completed", 00:16:06.469 "digest": "sha256", 00:16:06.469 "dhgroup": "ffdhe3072" 00:16:06.469 } 00:16:06.469 } 00:16:06.469 ]' 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:06.730 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:06.990 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:06.990 17:37:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:07.561 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:07.561 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 1 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:07.821 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:08.082 00:16:08.082 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:08.082 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:08.082 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:08.342 { 00:16:08.342 "cntlid": 19, 00:16:08.342 "qid": 0, 00:16:08.342 "state": "enabled", 00:16:08.342 "thread": "nvmf_tgt_poll_group_000", 00:16:08.342 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:08.342 "listen_address": { 00:16:08.342 "trtype": "TCP", 00:16:08.342 "adrfam": "IPv4", 00:16:08.342 "traddr": "10.0.0.2", 00:16:08.342 "trsvcid": "4420" 00:16:08.342 }, 00:16:08.342 "peer_address": { 00:16:08.342 "trtype": "TCP", 00:16:08.342 "adrfam": "IPv4", 00:16:08.342 "traddr": "10.0.0.1", 00:16:08.342 "trsvcid": "52116" 00:16:08.342 }, 00:16:08.342 "auth": { 00:16:08.342 "state": "completed", 00:16:08.342 "digest": "sha256", 00:16:08.342 "dhgroup": "ffdhe3072" 00:16:08.342 } 00:16:08.342 } 00:16:08.342 ]' 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:08.342 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:08.602 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:08.602 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:09.173 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:09.173 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:09.173 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:09.173 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.173 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:09.173 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.173 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:09.173 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:09.173 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 2 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:09.434 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:09.694 00:16:09.694 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:09.694 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:09.694 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:09.954 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:09.954 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:09.954 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.954 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:09.954 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:09.955 { 00:16:09.955 "cntlid": 21, 00:16:09.955 "qid": 0, 00:16:09.955 "state": "enabled", 00:16:09.955 "thread": "nvmf_tgt_poll_group_000", 00:16:09.955 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:09.955 "listen_address": { 00:16:09.955 "trtype": "TCP", 00:16:09.955 "adrfam": "IPv4", 00:16:09.955 "traddr": "10.0.0.2", 00:16:09.955 "trsvcid": "4420" 00:16:09.955 }, 00:16:09.955 "peer_address": { 00:16:09.955 "trtype": "TCP", 00:16:09.955 "adrfam": "IPv4", 00:16:09.955 "traddr": "10.0.0.1", 00:16:09.955 "trsvcid": "52126" 00:16:09.955 }, 00:16:09.955 "auth": { 00:16:09.955 "state": "completed", 00:16:09.955 "digest": "sha256", 00:16:09.955 "dhgroup": "ffdhe3072" 00:16:09.955 } 00:16:09.955 } 00:16:09.955 ]' 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:09.955 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:10.216 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:10.216 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:10.787 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:10.787 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 3 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:11.047 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:11.307 00:16:11.307 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:11.307 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:11.307 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:11.568 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:11.568 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:11.568 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.568 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:11.569 { 00:16:11.569 "cntlid": 23, 00:16:11.569 "qid": 0, 00:16:11.569 "state": "enabled", 00:16:11.569 "thread": "nvmf_tgt_poll_group_000", 00:16:11.569 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:11.569 "listen_address": { 00:16:11.569 "trtype": "TCP", 00:16:11.569 "adrfam": "IPv4", 00:16:11.569 "traddr": "10.0.0.2", 00:16:11.569 "trsvcid": "4420" 00:16:11.569 }, 00:16:11.569 "peer_address": { 00:16:11.569 "trtype": "TCP", 00:16:11.569 "adrfam": "IPv4", 00:16:11.569 "traddr": "10.0.0.1", 00:16:11.569 "trsvcid": "52166" 00:16:11.569 }, 00:16:11.569 "auth": { 00:16:11.569 "state": "completed", 00:16:11.569 "digest": "sha256", 00:16:11.569 "dhgroup": "ffdhe3072" 00:16:11.569 } 00:16:11.569 } 00:16:11.569 ]' 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:11.569 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:11.829 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:11.829 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:12.400 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:12.401 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:12.401 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 0 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:12.662 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:12.922 00:16:12.922 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:12.922 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:12.922 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:13.183 { 00:16:13.183 "cntlid": 25, 00:16:13.183 "qid": 0, 00:16:13.183 "state": "enabled", 00:16:13.183 "thread": "nvmf_tgt_poll_group_000", 00:16:13.183 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:13.183 "listen_address": { 00:16:13.183 "trtype": "TCP", 00:16:13.183 "adrfam": "IPv4", 00:16:13.183 "traddr": "10.0.0.2", 00:16:13.183 "trsvcid": "4420" 00:16:13.183 }, 00:16:13.183 "peer_address": { 00:16:13.183 "trtype": "TCP", 00:16:13.183 "adrfam": "IPv4", 00:16:13.183 "traddr": "10.0.0.1", 00:16:13.183 "trsvcid": "52182" 00:16:13.183 }, 00:16:13.183 "auth": { 00:16:13.183 "state": "completed", 00:16:13.183 "digest": "sha256", 00:16:13.183 "dhgroup": "ffdhe4096" 00:16:13.183 } 00:16:13.183 } 00:16:13.183 ]' 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:13.183 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:13.183 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:13.183 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:13.183 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:13.444 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:13.444 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:14.015 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:14.015 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 1 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.275 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:14.276 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.276 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:14.276 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:14.276 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:14.536 00:16:14.536 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:14.536 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:14.536 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:14.797 { 00:16:14.797 "cntlid": 27, 00:16:14.797 "qid": 0, 00:16:14.797 "state": "enabled", 00:16:14.797 "thread": "nvmf_tgt_poll_group_000", 00:16:14.797 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:14.797 "listen_address": { 00:16:14.797 "trtype": "TCP", 00:16:14.797 "adrfam": "IPv4", 00:16:14.797 "traddr": "10.0.0.2", 00:16:14.797 "trsvcid": "4420" 00:16:14.797 }, 00:16:14.797 "peer_address": { 00:16:14.797 "trtype": "TCP", 00:16:14.797 "adrfam": "IPv4", 00:16:14.797 "traddr": "10.0.0.1", 00:16:14.797 "trsvcid": "51086" 00:16:14.797 }, 00:16:14.797 "auth": { 00:16:14.797 "state": "completed", 00:16:14.797 "digest": "sha256", 00:16:14.797 "dhgroup": "ffdhe4096" 00:16:14.797 } 00:16:14.797 } 00:16:14.797 ]' 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:14.797 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:15.057 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:15.057 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:15.629 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:15.629 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 2 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:15.891 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:16.152 00:16:16.152 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:16.152 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:16.152 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:16.413 { 00:16:16.413 "cntlid": 29, 00:16:16.413 "qid": 0, 00:16:16.413 "state": "enabled", 00:16:16.413 "thread": "nvmf_tgt_poll_group_000", 00:16:16.413 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:16.413 "listen_address": { 00:16:16.413 "trtype": "TCP", 00:16:16.413 "adrfam": "IPv4", 00:16:16.413 "traddr": "10.0.0.2", 00:16:16.413 "trsvcid": "4420" 00:16:16.413 }, 00:16:16.413 "peer_address": { 00:16:16.413 "trtype": "TCP", 00:16:16.413 "adrfam": "IPv4", 00:16:16.413 "traddr": "10.0.0.1", 00:16:16.413 "trsvcid": "51112" 00:16:16.413 }, 00:16:16.413 "auth": { 00:16:16.413 "state": "completed", 00:16:16.413 "digest": "sha256", 00:16:16.413 "dhgroup": "ffdhe4096" 00:16:16.413 } 00:16:16.413 } 00:16:16.413 ]' 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:16.413 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:16.674 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:16.674 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:17.245 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:17.245 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:17.505 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 3 00:16:17.505 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:17.506 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:17.766 00:16:17.766 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:17.766 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:17.766 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:18.026 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:18.027 { 00:16:18.027 "cntlid": 31, 00:16:18.027 "qid": 0, 00:16:18.027 "state": "enabled", 00:16:18.027 "thread": "nvmf_tgt_poll_group_000", 00:16:18.027 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:18.027 "listen_address": { 00:16:18.027 "trtype": "TCP", 00:16:18.027 "adrfam": "IPv4", 00:16:18.027 "traddr": "10.0.0.2", 00:16:18.027 "trsvcid": "4420" 00:16:18.027 }, 00:16:18.027 "peer_address": { 00:16:18.027 "trtype": "TCP", 00:16:18.027 "adrfam": "IPv4", 00:16:18.027 "traddr": "10.0.0.1", 00:16:18.027 "trsvcid": "51154" 00:16:18.027 }, 00:16:18.027 "auth": { 00:16:18.027 "state": "completed", 00:16:18.027 "digest": "sha256", 00:16:18.027 "dhgroup": "ffdhe4096" 00:16:18.027 } 00:16:18.027 } 00:16:18.027 ]' 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:18.027 17:37:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:18.287 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:18.287 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:18.857 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:18.857 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 0 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:19.117 17:37:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:19.377 00:16:19.377 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:19.377 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:19.377 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:19.637 { 00:16:19.637 "cntlid": 33, 00:16:19.637 "qid": 0, 00:16:19.637 "state": "enabled", 00:16:19.637 "thread": "nvmf_tgt_poll_group_000", 00:16:19.637 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:19.637 "listen_address": { 00:16:19.637 "trtype": "TCP", 00:16:19.637 "adrfam": "IPv4", 00:16:19.637 "traddr": "10.0.0.2", 00:16:19.637 "trsvcid": "4420" 00:16:19.637 }, 00:16:19.637 "peer_address": { 00:16:19.637 "trtype": "TCP", 00:16:19.637 "adrfam": "IPv4", 00:16:19.637 "traddr": "10.0.0.1", 00:16:19.637 "trsvcid": "51184" 00:16:19.637 }, 00:16:19.637 "auth": { 00:16:19.637 "state": "completed", 00:16:19.637 "digest": "sha256", 00:16:19.637 "dhgroup": "ffdhe6144" 00:16:19.637 } 00:16:19.637 } 00:16:19.637 ]' 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:19.637 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:19.897 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:19.897 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:19.897 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:19.897 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:19.897 17:37:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:20.839 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 1 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:20.839 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:21.099 00:16:21.099 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:21.099 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:21.099 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:21.359 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:21.359 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:21.359 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.359 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:21.359 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.360 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:21.360 { 00:16:21.360 "cntlid": 35, 00:16:21.360 "qid": 0, 00:16:21.360 "state": "enabled", 00:16:21.360 "thread": "nvmf_tgt_poll_group_000", 00:16:21.360 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:21.360 "listen_address": { 00:16:21.360 "trtype": "TCP", 00:16:21.360 "adrfam": "IPv4", 00:16:21.360 "traddr": "10.0.0.2", 00:16:21.360 "trsvcid": "4420" 00:16:21.360 }, 00:16:21.360 "peer_address": { 00:16:21.360 "trtype": "TCP", 00:16:21.360 "adrfam": "IPv4", 00:16:21.360 "traddr": "10.0.0.1", 00:16:21.360 "trsvcid": "51210" 00:16:21.360 }, 00:16:21.360 "auth": { 00:16:21.360 "state": "completed", 00:16:21.360 "digest": "sha256", 00:16:21.360 "dhgroup": "ffdhe6144" 00:16:21.360 } 00:16:21.360 } 00:16:21.360 ]' 00:16:21.360 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:21.360 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:21.360 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:21.360 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:21.360 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:21.620 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:21.620 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:21.620 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:21.620 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:21.620 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:22.562 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:22.562 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:22.562 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 2 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:22.563 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:22.823 00:16:22.823 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:22.823 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:22.823 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:23.084 { 00:16:23.084 "cntlid": 37, 00:16:23.084 "qid": 0, 00:16:23.084 "state": "enabled", 00:16:23.084 "thread": "nvmf_tgt_poll_group_000", 00:16:23.084 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:23.084 "listen_address": { 00:16:23.084 "trtype": "TCP", 00:16:23.084 "adrfam": "IPv4", 00:16:23.084 "traddr": "10.0.0.2", 00:16:23.084 "trsvcid": "4420" 00:16:23.084 }, 00:16:23.084 "peer_address": { 00:16:23.084 "trtype": "TCP", 00:16:23.084 "adrfam": "IPv4", 00:16:23.084 "traddr": "10.0.0.1", 00:16:23.084 "trsvcid": "51230" 00:16:23.084 }, 00:16:23.084 "auth": { 00:16:23.084 "state": "completed", 00:16:23.084 "digest": "sha256", 00:16:23.084 "dhgroup": "ffdhe6144" 00:16:23.084 } 00:16:23.084 } 00:16:23.084 ]' 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:23.084 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:23.345 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:23.345 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:23.345 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:23.345 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:23.345 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:24.288 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:24.288 17:37:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 3 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:24.288 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:24.549 00:16:24.549 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:24.549 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:24.549 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:24.881 { 00:16:24.881 "cntlid": 39, 00:16:24.881 "qid": 0, 00:16:24.881 "state": "enabled", 00:16:24.881 "thread": "nvmf_tgt_poll_group_000", 00:16:24.881 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:24.881 "listen_address": { 00:16:24.881 "trtype": "TCP", 00:16:24.881 "adrfam": "IPv4", 00:16:24.881 "traddr": "10.0.0.2", 00:16:24.881 "trsvcid": "4420" 00:16:24.881 }, 00:16:24.881 "peer_address": { 00:16:24.881 "trtype": "TCP", 00:16:24.881 "adrfam": "IPv4", 00:16:24.881 "traddr": "10.0.0.1", 00:16:24.881 "trsvcid": "38856" 00:16:24.881 }, 00:16:24.881 "auth": { 00:16:24.881 "state": "completed", 00:16:24.881 "digest": "sha256", 00:16:24.881 "dhgroup": "ffdhe6144" 00:16:24.881 } 00:16:24.881 } 00:16:24.881 ]' 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:24.881 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:25.167 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:25.167 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:25.167 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:25.167 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:25.168 17:37:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:25.747 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:25.747 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 0 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:26.008 17:37:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:26.579 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:26.579 { 00:16:26.579 "cntlid": 41, 00:16:26.579 "qid": 0, 00:16:26.579 "state": "enabled", 00:16:26.579 "thread": "nvmf_tgt_poll_group_000", 00:16:26.579 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:26.579 "listen_address": { 00:16:26.579 "trtype": "TCP", 00:16:26.579 "adrfam": "IPv4", 00:16:26.579 "traddr": "10.0.0.2", 00:16:26.579 "trsvcid": "4420" 00:16:26.579 }, 00:16:26.579 "peer_address": { 00:16:26.579 "trtype": "TCP", 00:16:26.579 "adrfam": "IPv4", 00:16:26.579 "traddr": "10.0.0.1", 00:16:26.579 "trsvcid": "38876" 00:16:26.579 }, 00:16:26.579 "auth": { 00:16:26.579 "state": "completed", 00:16:26.579 "digest": "sha256", 00:16:26.579 "dhgroup": "ffdhe8192" 00:16:26.579 } 00:16:26.579 } 00:16:26.579 ]' 00:16:26.579 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:26.841 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:26.841 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:26.841 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:26.841 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:26.841 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:26.841 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:26.841 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:27.100 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:27.100 17:37:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:27.671 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:27.671 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 1 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:27.931 17:37:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:28.191 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:28.452 { 00:16:28.452 "cntlid": 43, 00:16:28.452 "qid": 0, 00:16:28.452 "state": "enabled", 00:16:28.452 "thread": "nvmf_tgt_poll_group_000", 00:16:28.452 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:28.452 "listen_address": { 00:16:28.452 "trtype": "TCP", 00:16:28.452 "adrfam": "IPv4", 00:16:28.452 "traddr": "10.0.0.2", 00:16:28.452 "trsvcid": "4420" 00:16:28.452 }, 00:16:28.452 "peer_address": { 00:16:28.452 "trtype": "TCP", 00:16:28.452 "adrfam": "IPv4", 00:16:28.452 "traddr": "10.0.0.1", 00:16:28.452 "trsvcid": "38902" 00:16:28.452 }, 00:16:28.452 "auth": { 00:16:28.452 "state": "completed", 00:16:28.452 "digest": "sha256", 00:16:28.452 "dhgroup": "ffdhe8192" 00:16:28.452 } 00:16:28.452 } 00:16:28.452 ]' 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:28.452 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:28.713 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:28.713 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:28.713 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:28.713 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:28.713 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:28.713 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:28.713 17:37:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:29.653 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 2 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:29.653 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:30.224 00:16:30.224 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:30.224 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:30.224 17:37:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:30.224 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:30.224 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:30.224 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.224 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:30.485 { 00:16:30.485 "cntlid": 45, 00:16:30.485 "qid": 0, 00:16:30.485 "state": "enabled", 00:16:30.485 "thread": "nvmf_tgt_poll_group_000", 00:16:30.485 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:30.485 "listen_address": { 00:16:30.485 "trtype": "TCP", 00:16:30.485 "adrfam": "IPv4", 00:16:30.485 "traddr": "10.0.0.2", 00:16:30.485 "trsvcid": "4420" 00:16:30.485 }, 00:16:30.485 "peer_address": { 00:16:30.485 "trtype": "TCP", 00:16:30.485 "adrfam": "IPv4", 00:16:30.485 "traddr": "10.0.0.1", 00:16:30.485 "trsvcid": "38924" 00:16:30.485 }, 00:16:30.485 "auth": { 00:16:30.485 "state": "completed", 00:16:30.485 "digest": "sha256", 00:16:30.485 "dhgroup": "ffdhe8192" 00:16:30.485 } 00:16:30.485 } 00:16:30.485 ]' 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:30.485 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:30.486 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:30.746 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:30.746 17:37:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:31.315 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:31.315 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 3 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:31.575 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:32.147 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:32.147 { 00:16:32.147 "cntlid": 47, 00:16:32.147 "qid": 0, 00:16:32.147 "state": "enabled", 00:16:32.147 "thread": "nvmf_tgt_poll_group_000", 00:16:32.147 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:32.147 "listen_address": { 00:16:32.147 "trtype": "TCP", 00:16:32.147 "adrfam": "IPv4", 00:16:32.147 "traddr": "10.0.0.2", 00:16:32.147 "trsvcid": "4420" 00:16:32.147 }, 00:16:32.147 "peer_address": { 00:16:32.147 "trtype": "TCP", 00:16:32.147 "adrfam": "IPv4", 00:16:32.147 "traddr": "10.0.0.1", 00:16:32.147 "trsvcid": "38940" 00:16:32.147 }, 00:16:32.147 "auth": { 00:16:32.147 "state": "completed", 00:16:32.147 "digest": "sha256", 00:16:32.147 "dhgroup": "ffdhe8192" 00:16:32.147 } 00:16:32.147 } 00:16:32.147 ]' 00:16:32.147 17:37:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:32.147 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:32.147 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:32.407 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:32.407 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:32.407 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:32.407 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:32.407 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:32.407 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:32.407 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:33.349 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:33.349 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:33.349 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:33.350 17:37:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 0 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:33.350 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:33.611 00:16:33.611 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:33.611 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:33.611 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:33.873 { 00:16:33.873 "cntlid": 49, 00:16:33.873 "qid": 0, 00:16:33.873 "state": "enabled", 00:16:33.873 "thread": "nvmf_tgt_poll_group_000", 00:16:33.873 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:33.873 "listen_address": { 00:16:33.873 "trtype": "TCP", 00:16:33.873 "adrfam": "IPv4", 00:16:33.873 "traddr": "10.0.0.2", 00:16:33.873 "trsvcid": "4420" 00:16:33.873 }, 00:16:33.873 "peer_address": { 00:16:33.873 "trtype": "TCP", 00:16:33.873 "adrfam": "IPv4", 00:16:33.873 "traddr": "10.0.0.1", 00:16:33.873 "trsvcid": "38960" 00:16:33.873 }, 00:16:33.873 "auth": { 00:16:33.873 "state": "completed", 00:16:33.873 "digest": "sha384", 00:16:33.873 "dhgroup": "null" 00:16:33.873 } 00:16:33.873 } 00:16:33.873 ]' 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:33.873 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:34.135 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:34.135 17:37:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:34.707 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:34.707 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 1 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:34.969 17:37:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:35.230 00:16:35.230 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:35.230 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:35.230 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:35.490 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:35.490 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:35.490 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.490 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:35.490 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.490 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:35.490 { 00:16:35.490 "cntlid": 51, 00:16:35.490 "qid": 0, 00:16:35.490 "state": "enabled", 00:16:35.490 "thread": "nvmf_tgt_poll_group_000", 00:16:35.490 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:35.490 "listen_address": { 00:16:35.491 "trtype": "TCP", 00:16:35.491 "adrfam": "IPv4", 00:16:35.491 "traddr": "10.0.0.2", 00:16:35.491 "trsvcid": "4420" 00:16:35.491 }, 00:16:35.491 "peer_address": { 00:16:35.491 "trtype": "TCP", 00:16:35.491 "adrfam": "IPv4", 00:16:35.491 "traddr": "10.0.0.1", 00:16:35.491 "trsvcid": "52812" 00:16:35.491 }, 00:16:35.491 "auth": { 00:16:35.491 "state": "completed", 00:16:35.491 "digest": "sha384", 00:16:35.491 "dhgroup": "null" 00:16:35.491 } 00:16:35.491 } 00:16:35.491 ]' 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:35.491 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:35.752 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:35.752 17:37:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:36.695 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 2 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:36.695 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:36.956 00:16:36.956 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:36.956 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:36.956 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:37.216 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:37.217 { 00:16:37.217 "cntlid": 53, 00:16:37.217 "qid": 0, 00:16:37.217 "state": "enabled", 00:16:37.217 "thread": "nvmf_tgt_poll_group_000", 00:16:37.217 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:37.217 "listen_address": { 00:16:37.217 "trtype": "TCP", 00:16:37.217 "adrfam": "IPv4", 00:16:37.217 "traddr": "10.0.0.2", 00:16:37.217 "trsvcid": "4420" 00:16:37.217 }, 00:16:37.217 "peer_address": { 00:16:37.217 "trtype": "TCP", 00:16:37.217 "adrfam": "IPv4", 00:16:37.217 "traddr": "10.0.0.1", 00:16:37.217 "trsvcid": "52828" 00:16:37.217 }, 00:16:37.217 "auth": { 00:16:37.217 "state": "completed", 00:16:37.217 "digest": "sha384", 00:16:37.217 "dhgroup": "null" 00:16:37.217 } 00:16:37.217 } 00:16:37.217 ]' 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:37.217 17:37:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:37.217 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:16:37.217 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:37.217 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:37.217 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:37.217 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:37.477 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:37.477 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:38.049 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:38.049 17:37:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 3 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:38.309 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:38.568 00:16:38.568 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:38.568 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:38.568 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:38.829 { 00:16:38.829 "cntlid": 55, 00:16:38.829 "qid": 0, 00:16:38.829 "state": "enabled", 00:16:38.829 "thread": "nvmf_tgt_poll_group_000", 00:16:38.829 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:38.829 "listen_address": { 00:16:38.829 "trtype": "TCP", 00:16:38.829 "adrfam": "IPv4", 00:16:38.829 "traddr": "10.0.0.2", 00:16:38.829 "trsvcid": "4420" 00:16:38.829 }, 00:16:38.829 "peer_address": { 00:16:38.829 "trtype": "TCP", 00:16:38.829 "adrfam": "IPv4", 00:16:38.829 "traddr": "10.0.0.1", 00:16:38.829 "trsvcid": "52848" 00:16:38.829 }, 00:16:38.829 "auth": { 00:16:38.829 "state": "completed", 00:16:38.829 "digest": "sha384", 00:16:38.829 "dhgroup": "null" 00:16:38.829 } 00:16:38.829 } 00:16:38.829 ]' 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:38.829 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:39.090 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:39.090 17:37:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:39.662 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:39.662 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:39.922 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 0 00:16:39.922 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:39.922 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:39.922 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:16:39.922 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:39.922 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:39.923 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:39.923 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.923 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:39.923 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.923 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:39.923 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:39.923 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:40.183 00:16:40.183 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:40.183 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:40.183 17:37:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:40.443 { 00:16:40.443 "cntlid": 57, 00:16:40.443 "qid": 0, 00:16:40.443 "state": "enabled", 00:16:40.443 "thread": "nvmf_tgt_poll_group_000", 00:16:40.443 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:40.443 "listen_address": { 00:16:40.443 "trtype": "TCP", 00:16:40.443 "adrfam": "IPv4", 00:16:40.443 "traddr": "10.0.0.2", 00:16:40.443 "trsvcid": "4420" 00:16:40.443 }, 00:16:40.443 "peer_address": { 00:16:40.443 "trtype": "TCP", 00:16:40.443 "adrfam": "IPv4", 00:16:40.443 "traddr": "10.0.0.1", 00:16:40.443 "trsvcid": "52874" 00:16:40.443 }, 00:16:40.443 "auth": { 00:16:40.443 "state": "completed", 00:16:40.443 "digest": "sha384", 00:16:40.443 "dhgroup": "ffdhe2048" 00:16:40.443 } 00:16:40.443 } 00:16:40.443 ]' 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:40.443 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:40.704 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:40.704 17:37:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:41.275 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:41.275 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 1 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:41.535 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:41.794 00:16:41.794 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:41.794 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:41.794 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:42.054 { 00:16:42.054 "cntlid": 59, 00:16:42.054 "qid": 0, 00:16:42.054 "state": "enabled", 00:16:42.054 "thread": "nvmf_tgt_poll_group_000", 00:16:42.054 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:42.054 "listen_address": { 00:16:42.054 "trtype": "TCP", 00:16:42.054 "adrfam": "IPv4", 00:16:42.054 "traddr": "10.0.0.2", 00:16:42.054 "trsvcid": "4420" 00:16:42.054 }, 00:16:42.054 "peer_address": { 00:16:42.054 "trtype": "TCP", 00:16:42.054 "adrfam": "IPv4", 00:16:42.054 "traddr": "10.0.0.1", 00:16:42.054 "trsvcid": "52906" 00:16:42.054 }, 00:16:42.054 "auth": { 00:16:42.054 "state": "completed", 00:16:42.054 "digest": "sha384", 00:16:42.054 "dhgroup": "ffdhe2048" 00:16:42.054 } 00:16:42.054 } 00:16:42.054 ]' 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:42.054 17:37:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:42.314 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:42.314 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:42.885 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:42.885 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:42.885 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:42.885 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.885 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 2 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:43.146 17:37:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:43.406 00:16:43.406 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:43.406 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:43.406 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:43.666 { 00:16:43.666 "cntlid": 61, 00:16:43.666 "qid": 0, 00:16:43.666 "state": "enabled", 00:16:43.666 "thread": "nvmf_tgt_poll_group_000", 00:16:43.666 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:43.666 "listen_address": { 00:16:43.666 "trtype": "TCP", 00:16:43.666 "adrfam": "IPv4", 00:16:43.666 "traddr": "10.0.0.2", 00:16:43.666 "trsvcid": "4420" 00:16:43.666 }, 00:16:43.666 "peer_address": { 00:16:43.666 "trtype": "TCP", 00:16:43.666 "adrfam": "IPv4", 00:16:43.666 "traddr": "10.0.0.1", 00:16:43.666 "trsvcid": "52940" 00:16:43.666 }, 00:16:43.666 "auth": { 00:16:43.666 "state": "completed", 00:16:43.666 "digest": "sha384", 00:16:43.666 "dhgroup": "ffdhe2048" 00:16:43.666 } 00:16:43.666 } 00:16:43.666 ]' 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:43.666 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:43.926 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:43.926 17:37:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:44.495 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:44.495 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 3 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:44.754 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:45.015 00:16:45.015 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:45.015 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:45.015 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:45.275 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:45.275 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:45.275 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.275 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:45.275 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.275 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:45.275 { 00:16:45.275 "cntlid": 63, 00:16:45.275 "qid": 0, 00:16:45.275 "state": "enabled", 00:16:45.275 "thread": "nvmf_tgt_poll_group_000", 00:16:45.275 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:45.275 "listen_address": { 00:16:45.275 "trtype": "TCP", 00:16:45.275 "adrfam": "IPv4", 00:16:45.275 "traddr": "10.0.0.2", 00:16:45.275 "trsvcid": "4420" 00:16:45.275 }, 00:16:45.275 "peer_address": { 00:16:45.275 "trtype": "TCP", 00:16:45.275 "adrfam": "IPv4", 00:16:45.275 "traddr": "10.0.0.1", 00:16:45.275 "trsvcid": "52746" 00:16:45.275 }, 00:16:45.275 "auth": { 00:16:45.275 "state": "completed", 00:16:45.275 "digest": "sha384", 00:16:45.275 "dhgroup": "ffdhe2048" 00:16:45.275 } 00:16:45.275 } 00:16:45.275 ]' 00:16:45.275 17:37:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:45.275 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:45.275 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:45.275 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:45.275 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:45.275 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:45.275 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:45.275 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:45.534 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:45.534 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:46.104 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:46.104 17:37:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 0 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:46.364 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.365 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:46.365 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:46.365 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:46.625 00:16:46.625 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:46.625 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:46.625 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:46.885 { 00:16:46.885 "cntlid": 65, 00:16:46.885 "qid": 0, 00:16:46.885 "state": "enabled", 00:16:46.885 "thread": "nvmf_tgt_poll_group_000", 00:16:46.885 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:46.885 "listen_address": { 00:16:46.885 "trtype": "TCP", 00:16:46.885 "adrfam": "IPv4", 00:16:46.885 "traddr": "10.0.0.2", 00:16:46.885 "trsvcid": "4420" 00:16:46.885 }, 00:16:46.885 "peer_address": { 00:16:46.885 "trtype": "TCP", 00:16:46.885 "adrfam": "IPv4", 00:16:46.885 "traddr": "10.0.0.1", 00:16:46.885 "trsvcid": "52766" 00:16:46.885 }, 00:16:46.885 "auth": { 00:16:46.885 "state": "completed", 00:16:46.885 "digest": "sha384", 00:16:46.885 "dhgroup": "ffdhe3072" 00:16:46.885 } 00:16:46.885 } 00:16:46.885 ]' 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:46.885 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:47.145 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:47.145 17:37:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:47.716 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:47.716 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 1 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:47.976 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:48.236 00:16:48.236 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:48.236 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:48.236 17:37:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:48.236 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:48.236 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:48.236 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.236 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:48.496 { 00:16:48.496 "cntlid": 67, 00:16:48.496 "qid": 0, 00:16:48.496 "state": "enabled", 00:16:48.496 "thread": "nvmf_tgt_poll_group_000", 00:16:48.496 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:48.496 "listen_address": { 00:16:48.496 "trtype": "TCP", 00:16:48.496 "adrfam": "IPv4", 00:16:48.496 "traddr": "10.0.0.2", 00:16:48.496 "trsvcid": "4420" 00:16:48.496 }, 00:16:48.496 "peer_address": { 00:16:48.496 "trtype": "TCP", 00:16:48.496 "adrfam": "IPv4", 00:16:48.496 "traddr": "10.0.0.1", 00:16:48.496 "trsvcid": "52788" 00:16:48.496 }, 00:16:48.496 "auth": { 00:16:48.496 "state": "completed", 00:16:48.496 "digest": "sha384", 00:16:48.496 "dhgroup": "ffdhe3072" 00:16:48.496 } 00:16:48.496 } 00:16:48.496 ]' 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:48.496 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:48.757 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:48.757 17:37:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:49.328 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:49.328 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 2 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:49.588 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:49.847 00:16:49.847 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:49.847 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:49.847 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:50.107 { 00:16:50.107 "cntlid": 69, 00:16:50.107 "qid": 0, 00:16:50.107 "state": "enabled", 00:16:50.107 "thread": "nvmf_tgt_poll_group_000", 00:16:50.107 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:50.107 "listen_address": { 00:16:50.107 "trtype": "TCP", 00:16:50.107 "adrfam": "IPv4", 00:16:50.107 "traddr": "10.0.0.2", 00:16:50.107 "trsvcid": "4420" 00:16:50.107 }, 00:16:50.107 "peer_address": { 00:16:50.107 "trtype": "TCP", 00:16:50.107 "adrfam": "IPv4", 00:16:50.107 "traddr": "10.0.0.1", 00:16:50.107 "trsvcid": "52814" 00:16:50.107 }, 00:16:50.107 "auth": { 00:16:50.107 "state": "completed", 00:16:50.107 "digest": "sha384", 00:16:50.107 "dhgroup": "ffdhe3072" 00:16:50.107 } 00:16:50.107 } 00:16:50.107 ]' 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:50.107 17:37:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:50.367 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:50.367 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:50.937 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:50.937 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:51.197 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 3 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:51.198 17:37:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:51.458 00:16:51.458 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:51.458 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:51.458 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:51.719 { 00:16:51.719 "cntlid": 71, 00:16:51.719 "qid": 0, 00:16:51.719 "state": "enabled", 00:16:51.719 "thread": "nvmf_tgt_poll_group_000", 00:16:51.719 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:51.719 "listen_address": { 00:16:51.719 "trtype": "TCP", 00:16:51.719 "adrfam": "IPv4", 00:16:51.719 "traddr": "10.0.0.2", 00:16:51.719 "trsvcid": "4420" 00:16:51.719 }, 00:16:51.719 "peer_address": { 00:16:51.719 "trtype": "TCP", 00:16:51.719 "adrfam": "IPv4", 00:16:51.719 "traddr": "10.0.0.1", 00:16:51.719 "trsvcid": "52830" 00:16:51.719 }, 00:16:51.719 "auth": { 00:16:51.719 "state": "completed", 00:16:51.719 "digest": "sha384", 00:16:51.719 "dhgroup": "ffdhe3072" 00:16:51.719 } 00:16:51.719 } 00:16:51.719 ]' 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:51.719 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:51.720 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:51.720 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:51.720 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:51.720 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:51.720 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:51.980 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:51.980 17:37:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:52.551 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:52.551 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 0 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:52.812 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:53.073 00:16:53.073 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:53.073 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:53.073 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:53.333 { 00:16:53.333 "cntlid": 73, 00:16:53.333 "qid": 0, 00:16:53.333 "state": "enabled", 00:16:53.333 "thread": "nvmf_tgt_poll_group_000", 00:16:53.333 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:53.333 "listen_address": { 00:16:53.333 "trtype": "TCP", 00:16:53.333 "adrfam": "IPv4", 00:16:53.333 "traddr": "10.0.0.2", 00:16:53.333 "trsvcid": "4420" 00:16:53.333 }, 00:16:53.333 "peer_address": { 00:16:53.333 "trtype": "TCP", 00:16:53.333 "adrfam": "IPv4", 00:16:53.333 "traddr": "10.0.0.1", 00:16:53.333 "trsvcid": "52864" 00:16:53.333 }, 00:16:53.333 "auth": { 00:16:53.333 "state": "completed", 00:16:53.333 "digest": "sha384", 00:16:53.333 "dhgroup": "ffdhe4096" 00:16:53.333 } 00:16:53.333 } 00:16:53.333 ]' 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:53.333 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:53.593 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:53.593 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:16:54.163 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:54.163 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:54.163 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:54.163 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.163 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:54.163 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.163 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:54.163 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:54.163 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 1 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:54.423 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:54.683 00:16:54.683 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:54.683 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:54.683 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:54.944 { 00:16:54.944 "cntlid": 75, 00:16:54.944 "qid": 0, 00:16:54.944 "state": "enabled", 00:16:54.944 "thread": "nvmf_tgt_poll_group_000", 00:16:54.944 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:54.944 "listen_address": { 00:16:54.944 "trtype": "TCP", 00:16:54.944 "adrfam": "IPv4", 00:16:54.944 "traddr": "10.0.0.2", 00:16:54.944 "trsvcid": "4420" 00:16:54.944 }, 00:16:54.944 "peer_address": { 00:16:54.944 "trtype": "TCP", 00:16:54.944 "adrfam": "IPv4", 00:16:54.944 "traddr": "10.0.0.1", 00:16:54.944 "trsvcid": "34070" 00:16:54.944 }, 00:16:54.944 "auth": { 00:16:54.944 "state": "completed", 00:16:54.944 "digest": "sha384", 00:16:54.944 "dhgroup": "ffdhe4096" 00:16:54.944 } 00:16:54.944 } 00:16:54.944 ]' 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:54.944 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:55.204 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:55.204 17:38:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:55.775 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:55.775 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 2 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:56.036 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:56.297 00:16:56.297 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:56.297 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:56.297 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:56.557 { 00:16:56.557 "cntlid": 77, 00:16:56.557 "qid": 0, 00:16:56.557 "state": "enabled", 00:16:56.557 "thread": "nvmf_tgt_poll_group_000", 00:16:56.557 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:56.557 "listen_address": { 00:16:56.557 "trtype": "TCP", 00:16:56.557 "adrfam": "IPv4", 00:16:56.557 "traddr": "10.0.0.2", 00:16:56.557 "trsvcid": "4420" 00:16:56.557 }, 00:16:56.557 "peer_address": { 00:16:56.557 "trtype": "TCP", 00:16:56.557 "adrfam": "IPv4", 00:16:56.557 "traddr": "10.0.0.1", 00:16:56.557 "trsvcid": "34106" 00:16:56.557 }, 00:16:56.557 "auth": { 00:16:56.557 "state": "completed", 00:16:56.557 "digest": "sha384", 00:16:56.557 "dhgroup": "ffdhe4096" 00:16:56.557 } 00:16:56.557 } 00:16:56.557 ]' 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:56.557 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:56.818 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:56.818 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:57.389 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:57.389 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:57.650 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 3 00:16:57.650 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:57.651 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:16:57.911 00:16:57.911 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:57.911 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:57.911 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:58.172 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:58.173 { 00:16:58.173 "cntlid": 79, 00:16:58.173 "qid": 0, 00:16:58.173 "state": "enabled", 00:16:58.173 "thread": "nvmf_tgt_poll_group_000", 00:16:58.173 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:58.173 "listen_address": { 00:16:58.173 "trtype": "TCP", 00:16:58.173 "adrfam": "IPv4", 00:16:58.173 "traddr": "10.0.0.2", 00:16:58.173 "trsvcid": "4420" 00:16:58.173 }, 00:16:58.173 "peer_address": { 00:16:58.173 "trtype": "TCP", 00:16:58.173 "adrfam": "IPv4", 00:16:58.173 "traddr": "10.0.0.1", 00:16:58.173 "trsvcid": "34138" 00:16:58.173 }, 00:16:58.173 "auth": { 00:16:58.173 "state": "completed", 00:16:58.173 "digest": "sha384", 00:16:58.173 "dhgroup": "ffdhe4096" 00:16:58.173 } 00:16:58.173 } 00:16:58.173 ]' 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:58.173 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:58.173 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:58.173 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:58.173 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:58.173 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:58.173 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:58.434 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:58.434 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:59.004 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:16:59.004 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 0 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:59.264 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:59.528 00:16:59.528 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:16:59.528 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:16:59.528 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:59.788 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:59.788 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:59.788 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.788 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:59.788 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.788 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:16:59.788 { 00:16:59.788 "cntlid": 81, 00:16:59.788 "qid": 0, 00:16:59.788 "state": "enabled", 00:16:59.788 "thread": "nvmf_tgt_poll_group_000", 00:16:59.788 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:16:59.789 "listen_address": { 00:16:59.789 "trtype": "TCP", 00:16:59.789 "adrfam": "IPv4", 00:16:59.789 "traddr": "10.0.0.2", 00:16:59.789 "trsvcid": "4420" 00:16:59.789 }, 00:16:59.789 "peer_address": { 00:16:59.789 "trtype": "TCP", 00:16:59.789 "adrfam": "IPv4", 00:16:59.789 "traddr": "10.0.0.1", 00:16:59.789 "trsvcid": "34160" 00:16:59.789 }, 00:16:59.789 "auth": { 00:16:59.789 "state": "completed", 00:16:59.789 "digest": "sha384", 00:16:59.789 "dhgroup": "ffdhe6144" 00:16:59.789 } 00:16:59.789 } 00:16:59.789 ]' 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:59.789 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:00.049 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:00.049 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:00.620 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:00.880 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 1 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.880 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:00.881 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:00.881 17:38:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:01.452 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:01.452 { 00:17:01.452 "cntlid": 83, 00:17:01.452 "qid": 0, 00:17:01.452 "state": "enabled", 00:17:01.452 "thread": "nvmf_tgt_poll_group_000", 00:17:01.452 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:01.452 "listen_address": { 00:17:01.452 "trtype": "TCP", 00:17:01.452 "adrfam": "IPv4", 00:17:01.452 "traddr": "10.0.0.2", 00:17:01.452 "trsvcid": "4420" 00:17:01.452 }, 00:17:01.452 "peer_address": { 00:17:01.452 "trtype": "TCP", 00:17:01.452 "adrfam": "IPv4", 00:17:01.452 "traddr": "10.0.0.1", 00:17:01.452 "trsvcid": "34188" 00:17:01.452 }, 00:17:01.452 "auth": { 00:17:01.452 "state": "completed", 00:17:01.452 "digest": "sha384", 00:17:01.452 "dhgroup": "ffdhe6144" 00:17:01.452 } 00:17:01.452 } 00:17:01.452 ]' 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:01.452 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:01.713 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:01.713 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:01.713 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:01.713 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:01.713 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:01.713 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:01.713 17:38:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:02.658 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 2 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.658 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:02.917 00:17:03.178 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:03.178 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:03.178 17:38:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:03.178 { 00:17:03.178 "cntlid": 85, 00:17:03.178 "qid": 0, 00:17:03.178 "state": "enabled", 00:17:03.178 "thread": "nvmf_tgt_poll_group_000", 00:17:03.178 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:03.178 "listen_address": { 00:17:03.178 "trtype": "TCP", 00:17:03.178 "adrfam": "IPv4", 00:17:03.178 "traddr": "10.0.0.2", 00:17:03.178 "trsvcid": "4420" 00:17:03.178 }, 00:17:03.178 "peer_address": { 00:17:03.178 "trtype": "TCP", 00:17:03.178 "adrfam": "IPv4", 00:17:03.178 "traddr": "10.0.0.1", 00:17:03.178 "trsvcid": "34206" 00:17:03.178 }, 00:17:03.178 "auth": { 00:17:03.178 "state": "completed", 00:17:03.178 "digest": "sha384", 00:17:03.178 "dhgroup": "ffdhe6144" 00:17:03.178 } 00:17:03.178 } 00:17:03.178 ]' 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:03.178 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:03.470 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:03.470 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:03.470 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:03.470 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:03.470 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:03.470 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:03.470 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:04.123 17:38:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:04.123 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:04.123 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:04.123 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.123 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:04.123 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.123 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:04.123 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:04.123 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:04.384 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 3 00:17:04.384 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:04.384 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:04.385 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:04.644 00:17:04.644 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:04.644 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:04.644 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:04.905 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:04.905 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:04.905 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.905 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:04.905 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.905 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:04.905 { 00:17:04.905 "cntlid": 87, 00:17:04.905 "qid": 0, 00:17:04.905 "state": "enabled", 00:17:04.905 "thread": "nvmf_tgt_poll_group_000", 00:17:04.906 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:04.906 "listen_address": { 00:17:04.906 "trtype": "TCP", 00:17:04.906 "adrfam": "IPv4", 00:17:04.906 "traddr": "10.0.0.2", 00:17:04.906 "trsvcid": "4420" 00:17:04.906 }, 00:17:04.906 "peer_address": { 00:17:04.906 "trtype": "TCP", 00:17:04.906 "adrfam": "IPv4", 00:17:04.906 "traddr": "10.0.0.1", 00:17:04.906 "trsvcid": "52956" 00:17:04.906 }, 00:17:04.906 "auth": { 00:17:04.906 "state": "completed", 00:17:04.906 "digest": "sha384", 00:17:04.906 "dhgroup": "ffdhe6144" 00:17:04.906 } 00:17:04.906 } 00:17:04.906 ]' 00:17:04.906 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:04.906 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:04.906 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:04.906 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:04.906 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:05.166 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:05.166 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:05.166 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:05.166 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:05.166 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:06.108 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 0 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:06.108 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:06.678 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:06.678 { 00:17:06.678 "cntlid": 89, 00:17:06.678 "qid": 0, 00:17:06.678 "state": "enabled", 00:17:06.678 "thread": "nvmf_tgt_poll_group_000", 00:17:06.678 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:06.678 "listen_address": { 00:17:06.678 "trtype": "TCP", 00:17:06.678 "adrfam": "IPv4", 00:17:06.678 "traddr": "10.0.0.2", 00:17:06.678 "trsvcid": "4420" 00:17:06.678 }, 00:17:06.678 "peer_address": { 00:17:06.678 "trtype": "TCP", 00:17:06.678 "adrfam": "IPv4", 00:17:06.678 "traddr": "10.0.0.1", 00:17:06.678 "trsvcid": "52986" 00:17:06.678 }, 00:17:06.678 "auth": { 00:17:06.678 "state": "completed", 00:17:06.678 "digest": "sha384", 00:17:06.678 "dhgroup": "ffdhe8192" 00:17:06.678 } 00:17:06.678 } 00:17:06.678 ]' 00:17:06.678 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:06.938 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:06.938 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:06.938 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:06.938 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:06.939 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:06.939 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:06.939 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:07.198 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:07.198 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:07.768 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:07.769 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:07.769 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:07.769 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:07.769 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:07.769 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:07.769 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:07.769 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:07.769 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 1 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:08.029 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:08.030 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:08.030 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:08.030 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:08.290 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:08.551 { 00:17:08.551 "cntlid": 91, 00:17:08.551 "qid": 0, 00:17:08.551 "state": "enabled", 00:17:08.551 "thread": "nvmf_tgt_poll_group_000", 00:17:08.551 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:08.551 "listen_address": { 00:17:08.551 "trtype": "TCP", 00:17:08.551 "adrfam": "IPv4", 00:17:08.551 "traddr": "10.0.0.2", 00:17:08.551 "trsvcid": "4420" 00:17:08.551 }, 00:17:08.551 "peer_address": { 00:17:08.551 "trtype": "TCP", 00:17:08.551 "adrfam": "IPv4", 00:17:08.551 "traddr": "10.0.0.1", 00:17:08.551 "trsvcid": "53014" 00:17:08.551 }, 00:17:08.551 "auth": { 00:17:08.551 "state": "completed", 00:17:08.551 "digest": "sha384", 00:17:08.551 "dhgroup": "ffdhe8192" 00:17:08.551 } 00:17:08.551 } 00:17:08.551 ]' 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:08.551 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:08.812 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:08.812 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:08.812 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:08.812 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:08.812 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:09.072 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:09.072 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:09.642 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:09.642 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 2 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:09.903 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:10.164 00:17:10.164 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:10.164 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:10.164 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:10.425 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:10.425 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:10.425 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.425 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:10.425 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.425 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:10.425 { 00:17:10.425 "cntlid": 93, 00:17:10.425 "qid": 0, 00:17:10.425 "state": "enabled", 00:17:10.425 "thread": "nvmf_tgt_poll_group_000", 00:17:10.426 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:10.426 "listen_address": { 00:17:10.426 "trtype": "TCP", 00:17:10.426 "adrfam": "IPv4", 00:17:10.426 "traddr": "10.0.0.2", 00:17:10.426 "trsvcid": "4420" 00:17:10.426 }, 00:17:10.426 "peer_address": { 00:17:10.426 "trtype": "TCP", 00:17:10.426 "adrfam": "IPv4", 00:17:10.426 "traddr": "10.0.0.1", 00:17:10.426 "trsvcid": "53036" 00:17:10.426 }, 00:17:10.426 "auth": { 00:17:10.426 "state": "completed", 00:17:10.426 "digest": "sha384", 00:17:10.426 "dhgroup": "ffdhe8192" 00:17:10.426 } 00:17:10.426 } 00:17:10.426 ]' 00:17:10.426 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:10.426 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:10.426 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:10.687 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:10.687 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:10.687 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:10.687 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:10.687 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:10.687 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:10.687 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:11.624 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:11.624 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 3 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:11.625 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:12.196 00:17:12.196 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:12.196 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:12.196 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:12.196 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:12.196 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:12.196 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:12.196 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:12.196 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:12.196 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:12.196 { 00:17:12.196 "cntlid": 95, 00:17:12.196 "qid": 0, 00:17:12.196 "state": "enabled", 00:17:12.196 "thread": "nvmf_tgt_poll_group_000", 00:17:12.196 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:12.196 "listen_address": { 00:17:12.196 "trtype": "TCP", 00:17:12.196 "adrfam": "IPv4", 00:17:12.196 "traddr": "10.0.0.2", 00:17:12.196 "trsvcid": "4420" 00:17:12.196 }, 00:17:12.196 "peer_address": { 00:17:12.196 "trtype": "TCP", 00:17:12.196 "adrfam": "IPv4", 00:17:12.196 "traddr": "10.0.0.1", 00:17:12.196 "trsvcid": "53064" 00:17:12.196 }, 00:17:12.196 "auth": { 00:17:12.196 "state": "completed", 00:17:12.196 "digest": "sha384", 00:17:12.196 "dhgroup": "ffdhe8192" 00:17:12.196 } 00:17:12.196 } 00:17:12.196 ]' 00:17:12.196 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:12.457 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:12.457 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:12.457 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:12.457 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:12.457 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:12.457 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:12.457 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:12.717 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:12.717 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:13.287 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:13.287 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:13.288 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 0 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:13.548 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:13.808 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.808 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:13.808 { 00:17:13.808 "cntlid": 97, 00:17:13.808 "qid": 0, 00:17:13.808 "state": "enabled", 00:17:13.808 "thread": "nvmf_tgt_poll_group_000", 00:17:13.808 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:13.808 "listen_address": { 00:17:13.808 "trtype": "TCP", 00:17:13.808 "adrfam": "IPv4", 00:17:13.808 "traddr": "10.0.0.2", 00:17:13.808 "trsvcid": "4420" 00:17:13.808 }, 00:17:13.808 "peer_address": { 00:17:13.808 "trtype": "TCP", 00:17:13.809 "adrfam": "IPv4", 00:17:13.809 "traddr": "10.0.0.1", 00:17:13.809 "trsvcid": "53104" 00:17:13.809 }, 00:17:13.809 "auth": { 00:17:13.809 "state": "completed", 00:17:13.809 "digest": "sha512", 00:17:13.809 "dhgroup": "null" 00:17:13.809 } 00:17:13.809 } 00:17:13.809 ]' 00:17:13.809 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:14.070 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:14.070 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:14.070 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:14.070 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:14.070 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:14.070 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:14.071 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:14.332 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:14.332 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:14.904 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:14.904 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:14.904 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:14.904 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:14.904 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:14.905 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:14.905 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:14.905 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:14.905 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:15.165 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 1 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:15.166 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:15.166 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:15.427 { 00:17:15.427 "cntlid": 99, 00:17:15.427 "qid": 0, 00:17:15.427 "state": "enabled", 00:17:15.427 "thread": "nvmf_tgt_poll_group_000", 00:17:15.427 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:15.427 "listen_address": { 00:17:15.427 "trtype": "TCP", 00:17:15.427 "adrfam": "IPv4", 00:17:15.427 "traddr": "10.0.0.2", 00:17:15.427 "trsvcid": "4420" 00:17:15.427 }, 00:17:15.427 "peer_address": { 00:17:15.427 "trtype": "TCP", 00:17:15.427 "adrfam": "IPv4", 00:17:15.427 "traddr": "10.0.0.1", 00:17:15.427 "trsvcid": "44432" 00:17:15.427 }, 00:17:15.427 "auth": { 00:17:15.427 "state": "completed", 00:17:15.427 "digest": "sha512", 00:17:15.427 "dhgroup": "null" 00:17:15.427 } 00:17:15.427 } 00:17:15.427 ]' 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:15.427 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:15.688 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:15.688 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:15.688 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:15.688 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:15.688 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:15.949 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:15.949 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:16.521 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:16.521 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 2 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.782 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.783 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.783 00:17:17.044 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:17.044 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:17.044 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:17.045 { 00:17:17.045 "cntlid": 101, 00:17:17.045 "qid": 0, 00:17:17.045 "state": "enabled", 00:17:17.045 "thread": "nvmf_tgt_poll_group_000", 00:17:17.045 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:17.045 "listen_address": { 00:17:17.045 "trtype": "TCP", 00:17:17.045 "adrfam": "IPv4", 00:17:17.045 "traddr": "10.0.0.2", 00:17:17.045 "trsvcid": "4420" 00:17:17.045 }, 00:17:17.045 "peer_address": { 00:17:17.045 "trtype": "TCP", 00:17:17.045 "adrfam": "IPv4", 00:17:17.045 "traddr": "10.0.0.1", 00:17:17.045 "trsvcid": "44458" 00:17:17.045 }, 00:17:17.045 "auth": { 00:17:17.045 "state": "completed", 00:17:17.045 "digest": "sha512", 00:17:17.045 "dhgroup": "null" 00:17:17.045 } 00:17:17.045 } 00:17:17.045 ]' 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:17.045 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:17.305 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:17.305 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:17.305 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:17.305 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:17.305 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:17.567 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:17.567 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:18.139 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:18.139 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 3 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:18.399 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:18.659 00:17:18.659 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:18.659 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:18.659 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:18.659 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:18.660 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:18.660 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.660 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:18.660 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.660 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:18.660 { 00:17:18.660 "cntlid": 103, 00:17:18.660 "qid": 0, 00:17:18.660 "state": "enabled", 00:17:18.660 "thread": "nvmf_tgt_poll_group_000", 00:17:18.660 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:18.660 "listen_address": { 00:17:18.660 "trtype": "TCP", 00:17:18.660 "adrfam": "IPv4", 00:17:18.660 "traddr": "10.0.0.2", 00:17:18.660 "trsvcid": "4420" 00:17:18.660 }, 00:17:18.660 "peer_address": { 00:17:18.660 "trtype": "TCP", 00:17:18.660 "adrfam": "IPv4", 00:17:18.660 "traddr": "10.0.0.1", 00:17:18.660 "trsvcid": "44490" 00:17:18.660 }, 00:17:18.660 "auth": { 00:17:18.660 "state": "completed", 00:17:18.660 "digest": "sha512", 00:17:18.660 "dhgroup": "null" 00:17:18.660 } 00:17:18.660 } 00:17:18.660 ]' 00:17:18.660 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:18.921 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:18.921 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:18.921 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:17:18.921 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:18.921 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:18.921 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:18.921 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:19.182 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:19.182 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:19.751 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:19.751 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 0 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:20.011 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:20.011 00:17:20.271 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:20.271 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:20.272 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:20.272 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:20.272 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:20.272 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:20.272 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:20.272 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:20.272 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:20.272 { 00:17:20.272 "cntlid": 105, 00:17:20.272 "qid": 0, 00:17:20.272 "state": "enabled", 00:17:20.272 "thread": "nvmf_tgt_poll_group_000", 00:17:20.272 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:20.272 "listen_address": { 00:17:20.272 "trtype": "TCP", 00:17:20.272 "adrfam": "IPv4", 00:17:20.272 "traddr": "10.0.0.2", 00:17:20.272 "trsvcid": "4420" 00:17:20.272 }, 00:17:20.272 "peer_address": { 00:17:20.272 "trtype": "TCP", 00:17:20.272 "adrfam": "IPv4", 00:17:20.272 "traddr": "10.0.0.1", 00:17:20.272 "trsvcid": "44514" 00:17:20.272 }, 00:17:20.272 "auth": { 00:17:20.272 "state": "completed", 00:17:20.272 "digest": "sha512", 00:17:20.272 "dhgroup": "ffdhe2048" 00:17:20.272 } 00:17:20.272 } 00:17:20.272 ]' 00:17:20.272 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:20.531 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:20.531 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:20.531 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:20.531 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:20.531 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:20.531 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:20.531 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:20.791 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:20.791 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:21.363 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:21.363 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 1 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:21.623 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:21.623 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:21.883 { 00:17:21.883 "cntlid": 107, 00:17:21.883 "qid": 0, 00:17:21.883 "state": "enabled", 00:17:21.883 "thread": "nvmf_tgt_poll_group_000", 00:17:21.883 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:21.883 "listen_address": { 00:17:21.883 "trtype": "TCP", 00:17:21.883 "adrfam": "IPv4", 00:17:21.883 "traddr": "10.0.0.2", 00:17:21.883 "trsvcid": "4420" 00:17:21.883 }, 00:17:21.883 "peer_address": { 00:17:21.883 "trtype": "TCP", 00:17:21.883 "adrfam": "IPv4", 00:17:21.883 "traddr": "10.0.0.1", 00:17:21.883 "trsvcid": "44548" 00:17:21.883 }, 00:17:21.883 "auth": { 00:17:21.883 "state": "completed", 00:17:21.883 "digest": "sha512", 00:17:21.883 "dhgroup": "ffdhe2048" 00:17:21.883 } 00:17:21.883 } 00:17:21.883 ]' 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:21.883 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:22.143 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:22.143 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:22.143 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:22.143 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:22.143 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:22.143 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:22.143 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:23.083 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 2 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:23.083 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:23.343 00:17:23.343 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:23.343 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:23.343 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:23.603 { 00:17:23.603 "cntlid": 109, 00:17:23.603 "qid": 0, 00:17:23.603 "state": "enabled", 00:17:23.603 "thread": "nvmf_tgt_poll_group_000", 00:17:23.603 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:23.603 "listen_address": { 00:17:23.603 "trtype": "TCP", 00:17:23.603 "adrfam": "IPv4", 00:17:23.603 "traddr": "10.0.0.2", 00:17:23.603 "trsvcid": "4420" 00:17:23.603 }, 00:17:23.603 "peer_address": { 00:17:23.603 "trtype": "TCP", 00:17:23.603 "adrfam": "IPv4", 00:17:23.603 "traddr": "10.0.0.1", 00:17:23.603 "trsvcid": "44580" 00:17:23.603 }, 00:17:23.603 "auth": { 00:17:23.603 "state": "completed", 00:17:23.603 "digest": "sha512", 00:17:23.603 "dhgroup": "ffdhe2048" 00:17:23.603 } 00:17:23.603 } 00:17:23.603 ]' 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:23.603 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:23.864 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:23.864 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:24.434 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:24.434 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 3 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:24.695 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:24.955 00:17:24.955 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:24.955 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:24.955 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:25.215 { 00:17:25.215 "cntlid": 111, 00:17:25.215 "qid": 0, 00:17:25.215 "state": "enabled", 00:17:25.215 "thread": "nvmf_tgt_poll_group_000", 00:17:25.215 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:25.215 "listen_address": { 00:17:25.215 "trtype": "TCP", 00:17:25.215 "adrfam": "IPv4", 00:17:25.215 "traddr": "10.0.0.2", 00:17:25.215 "trsvcid": "4420" 00:17:25.215 }, 00:17:25.215 "peer_address": { 00:17:25.215 "trtype": "TCP", 00:17:25.215 "adrfam": "IPv4", 00:17:25.215 "traddr": "10.0.0.1", 00:17:25.215 "trsvcid": "38414" 00:17:25.215 }, 00:17:25.215 "auth": { 00:17:25.215 "state": "completed", 00:17:25.215 "digest": "sha512", 00:17:25.215 "dhgroup": "ffdhe2048" 00:17:25.215 } 00:17:25.215 } 00:17:25.215 ]' 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:25.215 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:25.215 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:25.215 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:25.215 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:25.215 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:25.215 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:25.475 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:25.475 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:26.045 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:26.045 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 0 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.305 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:26.565 00:17:26.565 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:26.565 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:26.565 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:26.825 { 00:17:26.825 "cntlid": 113, 00:17:26.825 "qid": 0, 00:17:26.825 "state": "enabled", 00:17:26.825 "thread": "nvmf_tgt_poll_group_000", 00:17:26.825 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:26.825 "listen_address": { 00:17:26.825 "trtype": "TCP", 00:17:26.825 "adrfam": "IPv4", 00:17:26.825 "traddr": "10.0.0.2", 00:17:26.825 "trsvcid": "4420" 00:17:26.825 }, 00:17:26.825 "peer_address": { 00:17:26.825 "trtype": "TCP", 00:17:26.825 "adrfam": "IPv4", 00:17:26.825 "traddr": "10.0.0.1", 00:17:26.825 "trsvcid": "38446" 00:17:26.825 }, 00:17:26.825 "auth": { 00:17:26.825 "state": "completed", 00:17:26.825 "digest": "sha512", 00:17:26.825 "dhgroup": "ffdhe3072" 00:17:26.825 } 00:17:26.825 } 00:17:26.825 ]' 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:26.825 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:27.085 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:27.086 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:27.656 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:27.656 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:27.916 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 1 00:17:27.916 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:27.916 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:27.916 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:27.916 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:27.916 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:27.916 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:27.917 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:27.917 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:27.917 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:27.917 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:27.917 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:27.917 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:28.177 00:17:28.177 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:28.177 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:28.177 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:28.436 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:28.436 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:28.437 { 00:17:28.437 "cntlid": 115, 00:17:28.437 "qid": 0, 00:17:28.437 "state": "enabled", 00:17:28.437 "thread": "nvmf_tgt_poll_group_000", 00:17:28.437 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:28.437 "listen_address": { 00:17:28.437 "trtype": "TCP", 00:17:28.437 "adrfam": "IPv4", 00:17:28.437 "traddr": "10.0.0.2", 00:17:28.437 "trsvcid": "4420" 00:17:28.437 }, 00:17:28.437 "peer_address": { 00:17:28.437 "trtype": "TCP", 00:17:28.437 "adrfam": "IPv4", 00:17:28.437 "traddr": "10.0.0.1", 00:17:28.437 "trsvcid": "38468" 00:17:28.437 }, 00:17:28.437 "auth": { 00:17:28.437 "state": "completed", 00:17:28.437 "digest": "sha512", 00:17:28.437 "dhgroup": "ffdhe3072" 00:17:28.437 } 00:17:28.437 } 00:17:28.437 ]' 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:28.437 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:28.697 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:28.697 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:29.266 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:29.266 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 2 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:29.528 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:29.788 00:17:29.789 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:29.789 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:29.789 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:30.049 { 00:17:30.049 "cntlid": 117, 00:17:30.049 "qid": 0, 00:17:30.049 "state": "enabled", 00:17:30.049 "thread": "nvmf_tgt_poll_group_000", 00:17:30.049 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:30.049 "listen_address": { 00:17:30.049 "trtype": "TCP", 00:17:30.049 "adrfam": "IPv4", 00:17:30.049 "traddr": "10.0.0.2", 00:17:30.049 "trsvcid": "4420" 00:17:30.049 }, 00:17:30.049 "peer_address": { 00:17:30.049 "trtype": "TCP", 00:17:30.049 "adrfam": "IPv4", 00:17:30.049 "traddr": "10.0.0.1", 00:17:30.049 "trsvcid": "38490" 00:17:30.049 }, 00:17:30.049 "auth": { 00:17:30.049 "state": "completed", 00:17:30.049 "digest": "sha512", 00:17:30.049 "dhgroup": "ffdhe3072" 00:17:30.049 } 00:17:30.049 } 00:17:30.049 ]' 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:30.049 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:30.309 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:30.309 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:30.879 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:31.139 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 3 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.139 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.139 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.139 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:31.139 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:31.139 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:31.399 00:17:31.399 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:31.399 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:31.399 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:31.660 { 00:17:31.660 "cntlid": 119, 00:17:31.660 "qid": 0, 00:17:31.660 "state": "enabled", 00:17:31.660 "thread": "nvmf_tgt_poll_group_000", 00:17:31.660 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:31.660 "listen_address": { 00:17:31.660 "trtype": "TCP", 00:17:31.660 "adrfam": "IPv4", 00:17:31.660 "traddr": "10.0.0.2", 00:17:31.660 "trsvcid": "4420" 00:17:31.660 }, 00:17:31.660 "peer_address": { 00:17:31.660 "trtype": "TCP", 00:17:31.660 "adrfam": "IPv4", 00:17:31.660 "traddr": "10.0.0.1", 00:17:31.660 "trsvcid": "38516" 00:17:31.660 }, 00:17:31.660 "auth": { 00:17:31.660 "state": "completed", 00:17:31.660 "digest": "sha512", 00:17:31.660 "dhgroup": "ffdhe3072" 00:17:31.660 } 00:17:31.660 } 00:17:31.660 ]' 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:31.660 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:31.921 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:31.921 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:31.921 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:31.921 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:31.921 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:32.491 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:32.491 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 0 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:32.752 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:33.011 00:17:33.011 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:33.011 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:33.011 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:33.271 { 00:17:33.271 "cntlid": 121, 00:17:33.271 "qid": 0, 00:17:33.271 "state": "enabled", 00:17:33.271 "thread": "nvmf_tgt_poll_group_000", 00:17:33.271 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:33.271 "listen_address": { 00:17:33.271 "trtype": "TCP", 00:17:33.271 "adrfam": "IPv4", 00:17:33.271 "traddr": "10.0.0.2", 00:17:33.271 "trsvcid": "4420" 00:17:33.271 }, 00:17:33.271 "peer_address": { 00:17:33.271 "trtype": "TCP", 00:17:33.271 "adrfam": "IPv4", 00:17:33.271 "traddr": "10.0.0.1", 00:17:33.271 "trsvcid": "38542" 00:17:33.271 }, 00:17:33.271 "auth": { 00:17:33.271 "state": "completed", 00:17:33.271 "digest": "sha512", 00:17:33.271 "dhgroup": "ffdhe4096" 00:17:33.271 } 00:17:33.271 } 00:17:33.271 ]' 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:33.271 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:33.531 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:33.531 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:33.531 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:33.531 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:33.531 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:34.102 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:34.363 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 1 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:34.363 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:34.623 00:17:34.623 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:34.623 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:34.623 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:34.884 { 00:17:34.884 "cntlid": 123, 00:17:34.884 "qid": 0, 00:17:34.884 "state": "enabled", 00:17:34.884 "thread": "nvmf_tgt_poll_group_000", 00:17:34.884 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:34.884 "listen_address": { 00:17:34.884 "trtype": "TCP", 00:17:34.884 "adrfam": "IPv4", 00:17:34.884 "traddr": "10.0.0.2", 00:17:34.884 "trsvcid": "4420" 00:17:34.884 }, 00:17:34.884 "peer_address": { 00:17:34.884 "trtype": "TCP", 00:17:34.884 "adrfam": "IPv4", 00:17:34.884 "traddr": "10.0.0.1", 00:17:34.884 "trsvcid": "48690" 00:17:34.884 }, 00:17:34.884 "auth": { 00:17:34.884 "state": "completed", 00:17:34.884 "digest": "sha512", 00:17:34.884 "dhgroup": "ffdhe4096" 00:17:34.884 } 00:17:34.884 } 00:17:34.884 ]' 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:34.884 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:35.144 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:35.144 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:35.144 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:35.144 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:35.144 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:35.714 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:35.975 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 2 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:35.975 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:36.235 00:17:36.236 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:36.236 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:36.236 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:36.496 { 00:17:36.496 "cntlid": 125, 00:17:36.496 "qid": 0, 00:17:36.496 "state": "enabled", 00:17:36.496 "thread": "nvmf_tgt_poll_group_000", 00:17:36.496 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:36.496 "listen_address": { 00:17:36.496 "trtype": "TCP", 00:17:36.496 "adrfam": "IPv4", 00:17:36.496 "traddr": "10.0.0.2", 00:17:36.496 "trsvcid": "4420" 00:17:36.496 }, 00:17:36.496 "peer_address": { 00:17:36.496 "trtype": "TCP", 00:17:36.496 "adrfam": "IPv4", 00:17:36.496 "traddr": "10.0.0.1", 00:17:36.496 "trsvcid": "48712" 00:17:36.496 }, 00:17:36.496 "auth": { 00:17:36.496 "state": "completed", 00:17:36.496 "digest": "sha512", 00:17:36.496 "dhgroup": "ffdhe4096" 00:17:36.496 } 00:17:36.496 } 00:17:36.496 ]' 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:36.496 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:36.757 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:36.757 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:36.757 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:36.757 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:36.757 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:37.327 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:37.588 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 3 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:37.588 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:37.849 00:17:37.849 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:37.849 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:37.849 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:38.109 { 00:17:38.109 "cntlid": 127, 00:17:38.109 "qid": 0, 00:17:38.109 "state": "enabled", 00:17:38.109 "thread": "nvmf_tgt_poll_group_000", 00:17:38.109 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:38.109 "listen_address": { 00:17:38.109 "trtype": "TCP", 00:17:38.109 "adrfam": "IPv4", 00:17:38.109 "traddr": "10.0.0.2", 00:17:38.109 "trsvcid": "4420" 00:17:38.109 }, 00:17:38.109 "peer_address": { 00:17:38.109 "trtype": "TCP", 00:17:38.109 "adrfam": "IPv4", 00:17:38.109 "traddr": "10.0.0.1", 00:17:38.109 "trsvcid": "48740" 00:17:38.109 }, 00:17:38.109 "auth": { 00:17:38.109 "state": "completed", 00:17:38.109 "digest": "sha512", 00:17:38.109 "dhgroup": "ffdhe4096" 00:17:38.109 } 00:17:38.109 } 00:17:38.109 ]' 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:38.109 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:38.369 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:38.369 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:38.369 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:38.369 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:38.369 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:38.937 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:38.937 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 0 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:39.197 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:39.455 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:39.717 { 00:17:39.717 "cntlid": 129, 00:17:39.717 "qid": 0, 00:17:39.717 "state": "enabled", 00:17:39.717 "thread": "nvmf_tgt_poll_group_000", 00:17:39.717 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:39.717 "listen_address": { 00:17:39.717 "trtype": "TCP", 00:17:39.717 "adrfam": "IPv4", 00:17:39.717 "traddr": "10.0.0.2", 00:17:39.717 "trsvcid": "4420" 00:17:39.717 }, 00:17:39.717 "peer_address": { 00:17:39.717 "trtype": "TCP", 00:17:39.717 "adrfam": "IPv4", 00:17:39.717 "traddr": "10.0.0.1", 00:17:39.717 "trsvcid": "48760" 00:17:39.717 }, 00:17:39.717 "auth": { 00:17:39.717 "state": "completed", 00:17:39.717 "digest": "sha512", 00:17:39.717 "dhgroup": "ffdhe6144" 00:17:39.717 } 00:17:39.717 } 00:17:39.717 ]' 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:39.717 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:39.977 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:39.977 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:39.977 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:39.977 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:39.977 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:39.977 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:39.977 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:40.916 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:40.916 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:40.916 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:40.916 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:40.916 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 1 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:40.917 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:41.177 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:41.456 { 00:17:41.456 "cntlid": 131, 00:17:41.456 "qid": 0, 00:17:41.456 "state": "enabled", 00:17:41.456 "thread": "nvmf_tgt_poll_group_000", 00:17:41.456 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:41.456 "listen_address": { 00:17:41.456 "trtype": "TCP", 00:17:41.456 "adrfam": "IPv4", 00:17:41.456 "traddr": "10.0.0.2", 00:17:41.456 "trsvcid": "4420" 00:17:41.456 }, 00:17:41.456 "peer_address": { 00:17:41.456 "trtype": "TCP", 00:17:41.456 "adrfam": "IPv4", 00:17:41.456 "traddr": "10.0.0.1", 00:17:41.456 "trsvcid": "48790" 00:17:41.456 }, 00:17:41.456 "auth": { 00:17:41.456 "state": "completed", 00:17:41.456 "digest": "sha512", 00:17:41.456 "dhgroup": "ffdhe6144" 00:17:41.456 } 00:17:41.456 } 00:17:41.456 ]' 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:41.456 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:41.804 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:41.804 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:41.804 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:41.804 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:41.804 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:41.804 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:41.804 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:42.454 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:42.454 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 2 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:42.715 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:42.976 00:17:42.976 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:42.976 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:42.976 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:43.237 { 00:17:43.237 "cntlid": 133, 00:17:43.237 "qid": 0, 00:17:43.237 "state": "enabled", 00:17:43.237 "thread": "nvmf_tgt_poll_group_000", 00:17:43.237 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:43.237 "listen_address": { 00:17:43.237 "trtype": "TCP", 00:17:43.237 "adrfam": "IPv4", 00:17:43.237 "traddr": "10.0.0.2", 00:17:43.237 "trsvcid": "4420" 00:17:43.237 }, 00:17:43.237 "peer_address": { 00:17:43.237 "trtype": "TCP", 00:17:43.237 "adrfam": "IPv4", 00:17:43.237 "traddr": "10.0.0.1", 00:17:43.237 "trsvcid": "48812" 00:17:43.237 }, 00:17:43.237 "auth": { 00:17:43.237 "state": "completed", 00:17:43.237 "digest": "sha512", 00:17:43.237 "dhgroup": "ffdhe6144" 00:17:43.237 } 00:17:43.237 } 00:17:43.237 ]' 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:43.237 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:43.497 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:43.497 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:43.497 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:43.497 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:43.497 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:44.067 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:44.328 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:44.328 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:44.328 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:44.328 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 3 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:44.328 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:44.898 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:44.898 { 00:17:44.898 "cntlid": 135, 00:17:44.898 "qid": 0, 00:17:44.898 "state": "enabled", 00:17:44.898 "thread": "nvmf_tgt_poll_group_000", 00:17:44.898 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:44.898 "listen_address": { 00:17:44.898 "trtype": "TCP", 00:17:44.898 "adrfam": "IPv4", 00:17:44.898 "traddr": "10.0.0.2", 00:17:44.898 "trsvcid": "4420" 00:17:44.898 }, 00:17:44.898 "peer_address": { 00:17:44.898 "trtype": "TCP", 00:17:44.898 "adrfam": "IPv4", 00:17:44.898 "traddr": "10.0.0.1", 00:17:44.898 "trsvcid": "54484" 00:17:44.898 }, 00:17:44.898 "auth": { 00:17:44.898 "state": "completed", 00:17:44.898 "digest": "sha512", 00:17:44.898 "dhgroup": "ffdhe6144" 00:17:44.898 } 00:17:44.898 } 00:17:44.898 ]' 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:44.898 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:45.158 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:45.158 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:45.158 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:45.158 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:45.158 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:45.158 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:45.158 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:46.098 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 0 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:46.098 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:46.668 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:46.668 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:46.668 { 00:17:46.668 "cntlid": 137, 00:17:46.668 "qid": 0, 00:17:46.668 "state": "enabled", 00:17:46.668 "thread": "nvmf_tgt_poll_group_000", 00:17:46.669 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:46.669 "listen_address": { 00:17:46.669 "trtype": "TCP", 00:17:46.669 "adrfam": "IPv4", 00:17:46.669 "traddr": "10.0.0.2", 00:17:46.669 "trsvcid": "4420" 00:17:46.669 }, 00:17:46.669 "peer_address": { 00:17:46.669 "trtype": "TCP", 00:17:46.669 "adrfam": "IPv4", 00:17:46.669 "traddr": "10.0.0.1", 00:17:46.669 "trsvcid": "54526" 00:17:46.669 }, 00:17:46.669 "auth": { 00:17:46.669 "state": "completed", 00:17:46.669 "digest": "sha512", 00:17:46.669 "dhgroup": "ffdhe8192" 00:17:46.669 } 00:17:46.669 } 00:17:46.669 ]' 00:17:46.669 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:46.669 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:46.669 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:46.930 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:46.930 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:46.930 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:46.930 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:46.930 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:46.930 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:46.930 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:47.870 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 1 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:47.870 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:48.439 00:17:48.439 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:48.439 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:48.440 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:48.699 { 00:17:48.699 "cntlid": 139, 00:17:48.699 "qid": 0, 00:17:48.699 "state": "enabled", 00:17:48.699 "thread": "nvmf_tgt_poll_group_000", 00:17:48.699 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:48.699 "listen_address": { 00:17:48.699 "trtype": "TCP", 00:17:48.699 "adrfam": "IPv4", 00:17:48.699 "traddr": "10.0.0.2", 00:17:48.699 "trsvcid": "4420" 00:17:48.699 }, 00:17:48.699 "peer_address": { 00:17:48.699 "trtype": "TCP", 00:17:48.699 "adrfam": "IPv4", 00:17:48.699 "traddr": "10.0.0.1", 00:17:48.699 "trsvcid": "54558" 00:17:48.699 }, 00:17:48.699 "auth": { 00:17:48.699 "state": "completed", 00:17:48.699 "digest": "sha512", 00:17:48.699 "dhgroup": "ffdhe8192" 00:17:48.699 } 00:17:48.699 } 00:17:48.699 ]' 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:48.699 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:48.960 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:48.960 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: --dhchap-ctrl-secret DHHC-1:02:NGIxM2ZhZjUyOGUyZWM2OWQ2ODQ1MWRkZTc5OWQ3YzU2MWQzMzQ0ZDFjMTU3ZDYzDkDmTw==: 00:17:49.538 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:49.538 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:49.538 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:49.538 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:49.538 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:49.538 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:49.539 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:49.539 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:49.539 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 2 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:49.798 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:50.367 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:50.367 { 00:17:50.367 "cntlid": 141, 00:17:50.367 "qid": 0, 00:17:50.367 "state": "enabled", 00:17:50.367 "thread": "nvmf_tgt_poll_group_000", 00:17:50.367 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:50.367 "listen_address": { 00:17:50.367 "trtype": "TCP", 00:17:50.367 "adrfam": "IPv4", 00:17:50.367 "traddr": "10.0.0.2", 00:17:50.367 "trsvcid": "4420" 00:17:50.367 }, 00:17:50.367 "peer_address": { 00:17:50.367 "trtype": "TCP", 00:17:50.367 "adrfam": "IPv4", 00:17:50.367 "traddr": "10.0.0.1", 00:17:50.367 "trsvcid": "54576" 00:17:50.367 }, 00:17:50.367 "auth": { 00:17:50.367 "state": "completed", 00:17:50.367 "digest": "sha512", 00:17:50.367 "dhgroup": "ffdhe8192" 00:17:50.367 } 00:17:50.367 } 00:17:50.367 ]' 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:50.367 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:50.627 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:50.627 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:50.627 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:50.627 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:50.627 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:50.627 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:50.627 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:01:MWM3Y2ZiNjM2MGY5ODg2Y2IyNGQxNGU4Y2U2OTY1NjL1IgOR: 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:51.566 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 3 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:51.566 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:52.136 00:17:52.136 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:52.136 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:52.136 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:52.136 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:52.136 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:52.136 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:52.136 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:52.136 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:52.136 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:52.136 { 00:17:52.136 "cntlid": 143, 00:17:52.136 "qid": 0, 00:17:52.136 "state": "enabled", 00:17:52.136 "thread": "nvmf_tgt_poll_group_000", 00:17:52.136 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:52.136 "listen_address": { 00:17:52.136 "trtype": "TCP", 00:17:52.136 "adrfam": "IPv4", 00:17:52.136 "traddr": "10.0.0.2", 00:17:52.136 "trsvcid": "4420" 00:17:52.136 }, 00:17:52.136 "peer_address": { 00:17:52.136 "trtype": "TCP", 00:17:52.136 "adrfam": "IPv4", 00:17:52.136 "traddr": "10.0.0.1", 00:17:52.136 "trsvcid": "54612" 00:17:52.136 }, 00:17:52.136 "auth": { 00:17:52.136 "state": "completed", 00:17:52.136 "digest": "sha512", 00:17:52.136 "dhgroup": "ffdhe8192" 00:17:52.136 } 00:17:52.136 } 00:17:52.136 ]' 00:17:52.136 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:52.397 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:52.397 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:52.397 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:52.397 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:52.397 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:52.397 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:52.397 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:52.657 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:52.657 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:53.228 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:53.229 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s sha256,sha384,sha512 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:53.229 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@141 -- # connect_authenticate sha512 ffdhe8192 0 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:53.489 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:54.059 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:54.060 { 00:17:54.060 "cntlid": 145, 00:17:54.060 "qid": 0, 00:17:54.060 "state": "enabled", 00:17:54.060 "thread": "nvmf_tgt_poll_group_000", 00:17:54.060 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:54.060 "listen_address": { 00:17:54.060 "trtype": "TCP", 00:17:54.060 "adrfam": "IPv4", 00:17:54.060 "traddr": "10.0.0.2", 00:17:54.060 "trsvcid": "4420" 00:17:54.060 }, 00:17:54.060 "peer_address": { 00:17:54.060 "trtype": "TCP", 00:17:54.060 "adrfam": "IPv4", 00:17:54.060 "traddr": "10.0.0.1", 00:17:54.060 "trsvcid": "54642" 00:17:54.060 }, 00:17:54.060 "auth": { 00:17:54.060 "state": "completed", 00:17:54.060 "digest": "sha512", 00:17:54.060 "dhgroup": "ffdhe8192" 00:17:54.060 } 00:17:54.060 } 00:17:54.060 ]' 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:54.060 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:54.320 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:54.320 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:54.320 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:54.320 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:54.320 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:54.580 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:54.580 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:00:NTViZTRiZTVlMjMxYmE5YjQ2MTIwMWI2OTRjODk4MmUzYjE4NjI4NDcyNjhlOTBhzt31/A==: --dhchap-ctrl-secret DHHC-1:03:NmIxZTU3ODZjZWZlNDk2MTRkMmQwYWE4MGY1NjkzYzY3MDIzYWMxNGQyYWNhMDJkMjY3NWVhYjIyNWQwOTRjNWVHo64=: 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:55.150 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@144 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@145 -- # NOT bdev_connect -b nvme0 --dhchap-key key2 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key2 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key2 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:17:55.150 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:17:55.719 request: 00:17:55.719 { 00:17:55.719 "name": "nvme0", 00:17:55.719 "trtype": "tcp", 00:17:55.719 "traddr": "10.0.0.2", 00:17:55.719 "adrfam": "ipv4", 00:17:55.719 "trsvcid": "4420", 00:17:55.719 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:55.719 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:55.719 "prchk_reftag": false, 00:17:55.719 "prchk_guard": false, 00:17:55.719 "hdgst": false, 00:17:55.719 "ddgst": false, 00:17:55.719 "dhchap_key": "key2", 00:17:55.719 "allow_unrecognized_csi": false, 00:17:55.719 "method": "bdev_nvme_attach_controller", 00:17:55.719 "req_id": 1 00:17:55.719 } 00:17:55.719 Got JSON-RPC error response 00:17:55.719 response: 00:17:55.719 { 00:17:55.719 "code": -5, 00:17:55.719 "message": "Input/output error" 00:17:55.719 } 00:17:55.719 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:17:55.719 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@146 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@149 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@150 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:55.720 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:55.979 request: 00:17:55.979 { 00:17:55.979 "name": "nvme0", 00:17:55.979 "trtype": "tcp", 00:17:55.979 "traddr": "10.0.0.2", 00:17:55.979 "adrfam": "ipv4", 00:17:55.979 "trsvcid": "4420", 00:17:55.979 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:55.979 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:55.979 "prchk_reftag": false, 00:17:55.979 "prchk_guard": false, 00:17:55.979 "hdgst": false, 00:17:55.979 "ddgst": false, 00:17:55.979 "dhchap_key": "key1", 00:17:55.979 "dhchap_ctrlr_key": "ckey2", 00:17:55.979 "allow_unrecognized_csi": false, 00:17:55.979 "method": "bdev_nvme_attach_controller", 00:17:55.979 "req_id": 1 00:17:55.979 } 00:17:55.979 Got JSON-RPC error response 00:17:55.979 response: 00:17:55.979 { 00:17:55.979 "code": -5, 00:17:55.979 "message": "Input/output error" 00:17:55.979 } 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@151 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@154 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@155 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:55.979 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:56.549 request: 00:17:56.549 { 00:17:56.549 "name": "nvme0", 00:17:56.549 "trtype": "tcp", 00:17:56.549 "traddr": "10.0.0.2", 00:17:56.549 "adrfam": "ipv4", 00:17:56.549 "trsvcid": "4420", 00:17:56.549 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:56.549 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:56.549 "prchk_reftag": false, 00:17:56.550 "prchk_guard": false, 00:17:56.550 "hdgst": false, 00:17:56.550 "ddgst": false, 00:17:56.550 "dhchap_key": "key1", 00:17:56.550 "dhchap_ctrlr_key": "ckey1", 00:17:56.550 "allow_unrecognized_csi": false, 00:17:56.550 "method": "bdev_nvme_attach_controller", 00:17:56.550 "req_id": 1 00:17:56.550 } 00:17:56.550 Got JSON-RPC error response 00:17:56.550 response: 00:17:56.550 { 00:17:56.550 "code": -5, 00:17:56.550 "message": "Input/output error" 00:17:56.550 } 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@156 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@159 -- # killprocess 19452 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@950 -- # '[' -z 19452 ']' 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # kill -0 19452 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # uname 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 19452 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 19452' 00:17:56.550 killing process with pid 19452 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@969 -- # kill 19452 00:17:56.550 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@974 -- # wait 19452 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@160 -- # nvmfappstart --wait-for-rpc -L nvmf_auth 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # nvmfpid=45294 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # waitforlisten 45294 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc -L nvmf_auth 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 45294 ']' 00:17:56.809 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:56.810 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:56.810 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:56.810 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:56.810 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@161 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@163 -- # waitforlisten 45294 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 45294 ']' 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:57.750 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@164 -- # rpc_cmd 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:57.750 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.750 null0 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.iNU 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha512.IWh ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.IWh 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.EzL 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha384.cFu ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.cFu 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.0Y8 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha256.VIR ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.VIR 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.RrJ 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n '' ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@179 -- # connect_authenticate sha512 ffdhe8192 3 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:58.010 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:17:58.580 nvme0n1 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:17:58.840 { 00:17:58.840 "cntlid": 1, 00:17:58.840 "qid": 0, 00:17:58.840 "state": "enabled", 00:17:58.840 "thread": "nvmf_tgt_poll_group_000", 00:17:58.840 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:17:58.840 "listen_address": { 00:17:58.840 "trtype": "TCP", 00:17:58.840 "adrfam": "IPv4", 00:17:58.840 "traddr": "10.0.0.2", 00:17:58.840 "trsvcid": "4420" 00:17:58.840 }, 00:17:58.840 "peer_address": { 00:17:58.840 "trtype": "TCP", 00:17:58.840 "adrfam": "IPv4", 00:17:58.840 "traddr": "10.0.0.1", 00:17:58.840 "trsvcid": "57458" 00:17:58.840 }, 00:17:58.840 "auth": { 00:17:58.840 "state": "completed", 00:17:58.840 "digest": "sha512", 00:17:58.840 "dhgroup": "ffdhe8192" 00:17:58.840 } 00:17:58.840 } 00:17:58.840 ]' 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:58.840 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:17:59.100 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:59.100 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:17:59.100 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:59.100 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:59.100 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:59.100 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:17:59.100 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:00.039 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@182 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key3 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@183 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@184 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:00.039 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:00.299 request: 00:18:00.299 { 00:18:00.299 "name": "nvme0", 00:18:00.299 "trtype": "tcp", 00:18:00.299 "traddr": "10.0.0.2", 00:18:00.299 "adrfam": "ipv4", 00:18:00.299 "trsvcid": "4420", 00:18:00.299 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:18:00.299 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:18:00.299 "prchk_reftag": false, 00:18:00.299 "prchk_guard": false, 00:18:00.299 "hdgst": false, 00:18:00.299 "ddgst": false, 00:18:00.299 "dhchap_key": "key3", 00:18:00.299 "allow_unrecognized_csi": false, 00:18:00.299 "method": "bdev_nvme_attach_controller", 00:18:00.299 "req_id": 1 00:18:00.299 } 00:18:00.299 Got JSON-RPC error response 00:18:00.299 response: 00:18:00.299 { 00:18:00.299 "code": -5, 00:18:00.299 "message": "Input/output error" 00:18:00.299 } 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # IFS=, 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@188 -- # printf %s sha256,sha384,sha512 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # hostrpc bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:18:00.299 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@193 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key3 00:18:00.559 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:00.560 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:18:00.560 request: 00:18:00.560 { 00:18:00.560 "name": "nvme0", 00:18:00.560 "trtype": "tcp", 00:18:00.560 "traddr": "10.0.0.2", 00:18:00.560 "adrfam": "ipv4", 00:18:00.560 "trsvcid": "4420", 00:18:00.560 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:18:00.560 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:18:00.560 "prchk_reftag": false, 00:18:00.560 "prchk_guard": false, 00:18:00.560 "hdgst": false, 00:18:00.560 "ddgst": false, 00:18:00.560 "dhchap_key": "key3", 00:18:00.560 "allow_unrecognized_csi": false, 00:18:00.560 "method": "bdev_nvme_attach_controller", 00:18:00.560 "req_id": 1 00:18:00.560 } 00:18:00.560 Got JSON-RPC error response 00:18:00.560 response: 00:18:00.560 { 00:18:00.560 "code": -5, 00:18:00.560 "message": "Input/output error" 00:18:00.560 } 00:18:00.560 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:18:00.560 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:00.560 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:00.560 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:00.560 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:18:00.560 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s sha256,sha384,sha512 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@208 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@209 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@210 -- # NOT bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:18:00.820 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:18:01.080 request: 00:18:01.080 { 00:18:01.080 "name": "nvme0", 00:18:01.080 "trtype": "tcp", 00:18:01.080 "traddr": "10.0.0.2", 00:18:01.080 "adrfam": "ipv4", 00:18:01.080 "trsvcid": "4420", 00:18:01.080 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:18:01.080 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:18:01.080 "prchk_reftag": false, 00:18:01.080 "prchk_guard": false, 00:18:01.080 "hdgst": false, 00:18:01.080 "ddgst": false, 00:18:01.080 "dhchap_key": "key0", 00:18:01.080 "dhchap_ctrlr_key": "key1", 00:18:01.080 "allow_unrecognized_csi": false, 00:18:01.080 "method": "bdev_nvme_attach_controller", 00:18:01.080 "req_id": 1 00:18:01.080 } 00:18:01.080 Got JSON-RPC error response 00:18:01.080 response: 00:18:01.080 { 00:18:01.080 "code": -5, 00:18:01.080 "message": "Input/output error" 00:18:01.080 } 00:18:01.080 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:18:01.080 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:01.080 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:01.080 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:01.080 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@213 -- # bdev_connect -b nvme0 --dhchap-key key0 00:18:01.080 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:18:01.080 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:18:01.339 nvme0n1 00:18:01.339 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # hostrpc bdev_nvme_get_controllers 00:18:01.339 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # jq -r '.[].name' 00:18:01.339 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:01.599 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:01.599 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@215 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:01.599 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:01.859 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@218 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 00:18:01.859 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:01.859 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:01.859 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:01.859 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@219 -- # bdev_connect -b nvme0 --dhchap-key key1 00:18:01.859 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:18:01.859 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:18:02.429 nvme0n1 00:18:02.429 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # hostrpc bdev_nvme_get_controllers 00:18:02.429 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # jq -r '.[].name' 00:18:02.429 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@222 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # hostrpc bdev_nvme_get_controllers 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # jq -r '.[].name' 00:18:02.688 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:02.949 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:02.949 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@225 -- # nvme_connect --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:18:02.949 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid 008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -l 0 --dhchap-secret DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: --dhchap-ctrl-secret DHHC-1:03:ZWFhNzBhNDEyZmYyMDhkMDU2ZWU4NzhkZWZlNWZhMDUyOTM5NjUxNzU5YWM2YjA5YWJmYzJiMjE0NzA1MGFiYZTEZAw=: 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nvme_get_ctrlr 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@41 -- # local dev 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@43 -- # for dev in /sys/devices/virtual/nvme-fabrics/ctl/nvme* 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # [[ nqn.2024-03.io.spdk:cnode0 == \n\q\n\.\2\0\2\4\-\0\3\.\i\o\.\s\p\d\k\:\c\n\o\d\e\0 ]] 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # echo nvme0 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # break 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nctrlr=nvme0 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@227 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:03.519 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@228 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key1 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:18:03.779 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:18:04.041 request: 00:18:04.041 { 00:18:04.041 "name": "nvme0", 00:18:04.041 "trtype": "tcp", 00:18:04.041 "traddr": "10.0.0.2", 00:18:04.041 "adrfam": "ipv4", 00:18:04.041 "trsvcid": "4420", 00:18:04.041 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:18:04.041 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6", 00:18:04.041 "prchk_reftag": false, 00:18:04.041 "prchk_guard": false, 00:18:04.041 "hdgst": false, 00:18:04.041 "ddgst": false, 00:18:04.041 "dhchap_key": "key1", 00:18:04.041 "allow_unrecognized_csi": false, 00:18:04.041 "method": "bdev_nvme_attach_controller", 00:18:04.041 "req_id": 1 00:18:04.041 } 00:18:04.041 Got JSON-RPC error response 00:18:04.041 response: 00:18:04.041 { 00:18:04.041 "code": -5, 00:18:04.041 "message": "Input/output error" 00:18:04.041 } 00:18:04.301 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:18:04.301 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:04.301 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:04.301 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:04.301 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@229 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:04.301 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:04.301 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:04.870 nvme0n1 00:18:04.870 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # hostrpc bdev_nvme_get_controllers 00:18:04.870 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # jq -r '.[].name' 00:18:04.870 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:05.129 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:05.129 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@231 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:05.129 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@233 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@234 -- # bdev_connect -b nvme0 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:18:05.390 nvme0n1 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # hostrpc bdev_nvme_get_controllers 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:05.390 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # jq -r '.[].name' 00:18:05.652 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:05.652 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@236 -- # hostrpc bdev_nvme_detach_controller nvme0 00:18:05.652 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@239 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key key3 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@240 -- # nvme_set_keys nvme0 DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: '' 2s 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key=DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey= 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: ]] 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # echo DHHC-1:01:OGIyMDRiM2NiOTdmOGU4MjBkOGNiMjFlZjRiYmQ3NTju2A9H: 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z '' ]] 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:18:05.912 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:18:07.825 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@241 -- # waitforblk nvme0n1 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1235 -- # local i=0 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # lsblk -l -o NAME 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # grep -q -w nvme0n1 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # lsblk -l -o NAME 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # grep -q -w nvme0n1 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # return 0 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@243 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key1 --dhchap-ctrlr-key key2 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@244 -- # nvme_set_keys nvme0 '' DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: 2s 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key= 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey=DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z '' ]] 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: ]] 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # echo DHHC-1:02:YjJmMGMxNGQ5NjI3NDkxNGE5ZDIzNWE1NjgwMGE4Y2Y0Yjg3Mjc2MGNjMDY1MjAwdcnJyg==: 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:18:07.826 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@245 -- # waitforblk nvme0n1 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1235 -- # local i=0 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # lsblk -l -o NAME 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # grep -q -w nvme0n1 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # lsblk -l -o NAME 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # grep -q -w nvme0n1 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # return 0 00:18:10.447 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@246 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:18:10.447 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:18:10.448 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@249 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key key1 00:18:10.448 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:10.448 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:10.448 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:10.448 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@250 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:18:10.448 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:18:10.448 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:18:10.709 nvme0n1 00:18:10.709 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@252 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:10.709 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:10.709 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:10.709 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:10.709 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@253 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:10.709 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:11.278 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # hostrpc bdev_nvme_get_controllers 00:18:11.278 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # jq -r '.[].name' 00:18:11.278 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@256 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@257 -- # hostrpc bdev_nvme_set_keys nvme0 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # hostrpc bdev_nvme_get_controllers 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:11.539 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # jq -r '.[].name' 00:18:11.799 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:18:11.799 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@260 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@261 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=hostrpc 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t hostrpc 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:18:11.800 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:18:12.370 request: 00:18:12.370 { 00:18:12.370 "name": "nvme0", 00:18:12.370 "dhchap_key": "key1", 00:18:12.370 "dhchap_ctrlr_key": "key3", 00:18:12.370 "method": "bdev_nvme_set_keys", 00:18:12.370 "req_id": 1 00:18:12.370 } 00:18:12.370 Got JSON-RPC error response 00:18:12.370 response: 00:18:12.370 { 00:18:12.370 "code": -13, 00:18:12.370 "message": "Permission denied" 00:18:12.370 } 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 1 != 0 )) 00:18:12.370 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@263 -- # sleep 1s 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 0 != 0 )) 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@267 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key0 --dhchap-ctrlr-key key1 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@268 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:18:13.752 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:18:14.324 nvme0n1 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@270 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --dhchap-key key2 --dhchap-ctrlr-key key3 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@271 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=hostrpc 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t hostrpc 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:18:14.324 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:18:14.895 request: 00:18:14.895 { 00:18:14.895 "name": "nvme0", 00:18:14.895 "dhchap_key": "key2", 00:18:14.895 "dhchap_ctrlr_key": "key0", 00:18:14.895 "method": "bdev_nvme_set_keys", 00:18:14.895 "req_id": 1 00:18:14.895 } 00:18:14.895 Got JSON-RPC error response 00:18:14.895 response: 00:18:14.895 { 00:18:14.895 "code": -13, 00:18:14.895 "message": "Permission denied" 00:18:14.895 } 00:18:14.895 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:18:14.895 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:14.895 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:14.895 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:14.895 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:18:14.895 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:18:14.895 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:15.156 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 1 != 0 )) 00:18:15.156 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@273 -- # sleep 1s 00:18:16.096 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:18:16.096 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:18:16.096 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 0 != 0 )) 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@276 -- # trap - SIGINT SIGTERM EXIT 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@277 -- # cleanup 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@21 -- # killprocess 19666 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@950 -- # '[' -z 19666 ']' 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # kill -0 19666 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # uname 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 19666 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 19666' 00:18:16.357 killing process with pid 19666 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@969 -- # kill 19666 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@974 -- # wait 19666 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@22 -- # nvmftestfini 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:18:16.357 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@121 -- # sync 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@124 -- # set +e 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:18:16.617 rmmod nvme_tcp 00:18:16.617 rmmod nvme_fabrics 00:18:16.617 rmmod nvme_keyring 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@128 -- # set -e 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@129 -- # return 0 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@515 -- # '[' -n 45294 ']' 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@516 -- # killprocess 45294 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@950 -- # '[' -z 45294 ']' 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # kill -0 45294 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # uname 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 45294 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 45294' 00:18:16.617 killing process with pid 45294 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@969 -- # kill 45294 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@974 -- # wait 45294 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@297 -- # iptr 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@789 -- # iptables-save 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@789 -- # iptables-restore 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:18:16.617 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:19.157 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@23 -- # rm -f /tmp/spdk.key-null.iNU /tmp/spdk.key-sha256.EzL /tmp/spdk.key-sha384.0Y8 /tmp/spdk.key-sha512.RrJ /tmp/spdk.key-sha512.IWh /tmp/spdk.key-sha384.cFu /tmp/spdk.key-sha256.VIR '' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf-auth.log 00:18:19.158 00:18:19.158 real 2m36.833s 00:18:19.158 user 5m53.326s 00:18:19.158 sys 0m24.546s 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:19.158 ************************************ 00:18:19.158 END TEST nvmf_auth_target 00:18:19.158 ************************************ 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@39 -- # '[' tcp = tcp ']' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@40 -- # run_test nvmf_bdevio_no_huge /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:18:19.158 ************************************ 00:18:19.158 START TEST nvmf_bdevio_no_huge 00:18:19.158 ************************************ 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:18:19.158 * Looking for test storage... 00:18:19.158 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1691 -- # lcov --version 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # IFS=.-: 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # read -ra ver1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # IFS=.-: 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # read -ra ver2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@338 -- # local 'op=<' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@340 -- # ver1_l=2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@341 -- # ver2_l=1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@344 -- # case "$op" in 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@345 -- # : 1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # decimal 1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # ver1[v]=1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # decimal 2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # ver2[v]=2 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # return 0 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:18:19.158 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:19.158 --rc genhtml_branch_coverage=1 00:18:19.158 --rc genhtml_function_coverage=1 00:18:19.158 --rc genhtml_legend=1 00:18:19.158 --rc geninfo_all_blocks=1 00:18:19.158 --rc geninfo_unexecuted_blocks=1 00:18:19.158 00:18:19.158 ' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:18:19.158 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:19.158 --rc genhtml_branch_coverage=1 00:18:19.158 --rc genhtml_function_coverage=1 00:18:19.158 --rc genhtml_legend=1 00:18:19.158 --rc geninfo_all_blocks=1 00:18:19.158 --rc geninfo_unexecuted_blocks=1 00:18:19.158 00:18:19.158 ' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:18:19.158 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:19.158 --rc genhtml_branch_coverage=1 00:18:19.158 --rc genhtml_function_coverage=1 00:18:19.158 --rc genhtml_legend=1 00:18:19.158 --rc geninfo_all_blocks=1 00:18:19.158 --rc geninfo_unexecuted_blocks=1 00:18:19.158 00:18:19.158 ' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:18:19.158 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:19.158 --rc genhtml_branch_coverage=1 00:18:19.158 --rc genhtml_function_coverage=1 00:18:19.158 --rc genhtml_legend=1 00:18:19.158 --rc geninfo_all_blocks=1 00:18:19.158 --rc geninfo_unexecuted_blocks=1 00:18:19.158 00:18:19.158 ' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # uname -s 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@15 -- # shopt -s extglob 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:19.158 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@5 -- # export PATH 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@51 -- # : 0 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:18:19.159 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@55 -- # have_pci_nics=0 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@14 -- # nvmftestinit 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@474 -- # prepare_net_devs 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@436 -- # local -g is_hw=no 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@438 -- # remove_spdk_ns 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@309 -- # xtrace_disable 00:18:19.159 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.297 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:18:27.297 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # pci_devs=() 00:18:27.297 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # local -a pci_devs 00:18:27.297 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # pci_net_devs=() 00:18:27.297 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:18:27.297 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # pci_drivers=() 00:18:27.297 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # local -A pci_drivers 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # net_devs=() 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # local -ga net_devs 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # e810=() 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # local -ga e810 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # x722=() 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # local -ga x722 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # mlx=() 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # local -ga mlx 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:18:27.298 Found 0000:31:00.0 (0x8086 - 0x159b) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:18:27.298 Found 0000:31:00.1 (0x8086 - 0x159b) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ up == up ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:18:27.298 Found net devices under 0000:31:00.0: cvl_0_0 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ up == up ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:18:27.298 Found net devices under 0000:31:00.1: cvl_0_1 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # is_hw=yes 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:18:27.298 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:18:27.298 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.622 ms 00:18:27.298 00:18:27.298 --- 10.0.0.2 ping statistics --- 00:18:27.298 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:27.298 rtt min/avg/max/mdev = 0.622/0.622/0.622/0.000 ms 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:18:27.298 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:18:27.298 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.275 ms 00:18:27.298 00:18:27.298 --- 10.0.0.1 ping statistics --- 00:18:27.298 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:27.298 rtt min/avg/max/mdev = 0.275/0.275/0.275/0.000 ms 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@448 -- # return 0 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:18:27.298 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@724 -- # xtrace_disable 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@507 -- # nvmfpid=53943 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@508 -- # waitforlisten 53943 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --no-huge -s 1024 -m 0x78 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@831 -- # '[' -z 53943 ']' 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:27.299 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:27.299 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.299 [2024-10-17 17:39:34.610434] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:18:27.299 [2024-10-17 17:39:34.610502] [ DPDK EAL parameters: nvmf -c 0x78 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk0 --proc-type=auto ] 00:18:27.299 [2024-10-17 17:39:34.708187] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:18:27.299 [2024-10-17 17:39:34.766921] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:27.299 [2024-10-17 17:39:34.766964] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:27.299 [2024-10-17 17:39:34.766973] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:27.299 [2024-10-17 17:39:34.766979] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:27.299 [2024-10-17 17:39:34.766986] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:27.299 [2024-10-17 17:39:34.768518] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:18:27.299 [2024-10-17 17:39:34.768760] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:18:27.299 [2024-10-17 17:39:34.768993] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:18:27.299 [2024-10-17 17:39:34.769088] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:18:27.559 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:27.560 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@864 -- # return 0 00:18:27.560 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:18:27.560 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:27.560 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.820 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:27.820 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:18:27.820 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.821 [2024-10-17 17:39:35.492043] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.821 Malloc0 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:27.821 [2024-10-17 17:39:35.546023] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 --no-huge -s 1024 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@558 -- # config=() 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@558 -- # local subsystem config 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:18:27.821 { 00:18:27.821 "params": { 00:18:27.821 "name": "Nvme$subsystem", 00:18:27.821 "trtype": "$TEST_TRANSPORT", 00:18:27.821 "traddr": "$NVMF_FIRST_TARGET_IP", 00:18:27.821 "adrfam": "ipv4", 00:18:27.821 "trsvcid": "$NVMF_PORT", 00:18:27.821 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:18:27.821 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:18:27.821 "hdgst": ${hdgst:-false}, 00:18:27.821 "ddgst": ${ddgst:-false} 00:18:27.821 }, 00:18:27.821 "method": "bdev_nvme_attach_controller" 00:18:27.821 } 00:18:27.821 EOF 00:18:27.821 )") 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@580 -- # cat 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@582 -- # jq . 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@583 -- # IFS=, 00:18:27.821 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:18:27.821 "params": { 00:18:27.821 "name": "Nvme1", 00:18:27.821 "trtype": "tcp", 00:18:27.821 "traddr": "10.0.0.2", 00:18:27.821 "adrfam": "ipv4", 00:18:27.821 "trsvcid": "4420", 00:18:27.821 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:27.821 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:27.821 "hdgst": false, 00:18:27.821 "ddgst": false 00:18:27.821 }, 00:18:27.821 "method": "bdev_nvme_attach_controller" 00:18:27.821 }' 00:18:27.821 [2024-10-17 17:39:35.604002] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:18:27.821 [2024-10-17 17:39:35.604070] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk_pid54291 ] 00:18:27.821 [2024-10-17 17:39:35.691845] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:28.081 [2024-10-17 17:39:35.752265] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:28.081 [2024-10-17 17:39:35.752425] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:28.081 [2024-10-17 17:39:35.752425] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:18:28.341 I/O targets: 00:18:28.341 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:18:28.341 00:18:28.341 00:18:28.341 CUnit - A unit testing framework for C - Version 2.1-3 00:18:28.341 http://cunit.sourceforge.net/ 00:18:28.341 00:18:28.341 00:18:28.341 Suite: bdevio tests on: Nvme1n1 00:18:28.341 Test: blockdev write read block ...passed 00:18:28.341 Test: blockdev write zeroes read block ...passed 00:18:28.341 Test: blockdev write zeroes read no split ...passed 00:18:28.341 Test: blockdev write zeroes read split ...passed 00:18:28.341 Test: blockdev write zeroes read split partial ...passed 00:18:28.341 Test: blockdev reset ...[2024-10-17 17:39:36.199886] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:18:28.341 [2024-10-17 17:39:36.199984] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11d34f0 (9): Bad file descriptor 00:18:28.611 [2024-10-17 17:39:36.295921] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:18:28.611 passed 00:18:28.611 Test: blockdev write read 8 blocks ...passed 00:18:28.611 Test: blockdev write read size > 128k ...passed 00:18:28.611 Test: blockdev write read invalid size ...passed 00:18:28.611 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:18:28.611 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:18:28.611 Test: blockdev write read max offset ...passed 00:18:28.611 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:18:28.611 Test: blockdev writev readv 8 blocks ...passed 00:18:28.611 Test: blockdev writev readv 30 x 1block ...passed 00:18:28.611 Test: blockdev writev readv block ...passed 00:18:28.611 Test: blockdev writev readv size > 128k ...passed 00:18:28.611 Test: blockdev writev readv size > 128k in two iovs ...passed 00:18:28.611 Test: blockdev comparev and writev ...[2024-10-17 17:39:36.479971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.480020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:18:28.611 [2024-10-17 17:39:36.480038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.480047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:18:28.611 [2024-10-17 17:39:36.480652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.480666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:18:28.611 [2024-10-17 17:39:36.480681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.480689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:18:28.611 [2024-10-17 17:39:36.481252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.481267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:18:28.611 [2024-10-17 17:39:36.481282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.481289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:18:28.611 [2024-10-17 17:39:36.481860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.481884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:18:28.611 [2024-10-17 17:39:36.481897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:28.611 [2024-10-17 17:39:36.481912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:18:28.611 passed 00:18:28.984 Test: blockdev nvme passthru rw ...passed 00:18:28.984 Test: blockdev nvme passthru vendor specific ...[2024-10-17 17:39:36.566567] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:28.984 [2024-10-17 17:39:36.566587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:18:28.984 [2024-10-17 17:39:36.566975] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:28.984 [2024-10-17 17:39:36.566988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:18:28.984 [2024-10-17 17:39:36.567370] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:28.984 [2024-10-17 17:39:36.567383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:18:28.984 [2024-10-17 17:39:36.567770] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:28.984 [2024-10-17 17:39:36.567788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:18:28.984 passed 00:18:28.984 Test: blockdev nvme admin passthru ...passed 00:18:28.984 Test: blockdev copy ...passed 00:18:28.984 00:18:28.984 Run Summary: Type Total Ran Passed Failed Inactive 00:18:28.984 suites 1 1 n/a 0 0 00:18:28.984 tests 23 23 23 0 0 00:18:28.984 asserts 152 152 152 0 n/a 00:18:28.984 00:18:28.984 Elapsed time = 1.144 seconds 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@30 -- # nvmftestfini 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@514 -- # nvmfcleanup 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@121 -- # sync 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@124 -- # set +e 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@125 -- # for i in {1..20} 00:18:29.262 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:18:29.262 rmmod nvme_tcp 00:18:29.262 rmmod nvme_fabrics 00:18:29.262 rmmod nvme_keyring 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@128 -- # set -e 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@129 -- # return 0 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@515 -- # '[' -n 53943 ']' 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@516 -- # killprocess 53943 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@950 -- # '[' -z 53943 ']' 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@954 -- # kill -0 53943 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@955 -- # uname 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 53943 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@956 -- # process_name=reactor_3 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@960 -- # '[' reactor_3 = sudo ']' 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@968 -- # echo 'killing process with pid 53943' 00:18:29.262 killing process with pid 53943 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@969 -- # kill 53943 00:18:29.262 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@974 -- # wait 53943 00:18:29.536 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:18:29.536 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:18:29.536 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:18:29.536 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@297 -- # iptr 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@789 -- # iptables-save 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@789 -- # iptables-restore 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@302 -- # remove_spdk_ns 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:18:29.797 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:18:31.712 00:18:31.712 real 0m12.864s 00:18:31.712 user 0m15.120s 00:18:31.712 sys 0m6.817s 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:31.712 ************************************ 00:18:31.712 END TEST nvmf_bdevio_no_huge 00:18:31.712 ************************************ 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@41 -- # run_test nvmf_tls /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:18:31.712 ************************************ 00:18:31.712 START TEST nvmf_tls 00:18:31.712 ************************************ 00:18:31.712 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:18:31.974 * Looking for test storage... 00:18:31.974 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1691 -- # lcov --version 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # IFS=.-: 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # read -ra ver1 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # IFS=.-: 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # read -ra ver2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@338 -- # local 'op=<' 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@340 -- # ver1_l=2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@341 -- # ver2_l=1 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@344 -- # case "$op" in 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@345 -- # : 1 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # decimal 1 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=1 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 1 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # ver1[v]=1 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # decimal 2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # ver2[v]=2 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # return 0 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:18:31.974 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:31.974 --rc genhtml_branch_coverage=1 00:18:31.974 --rc genhtml_function_coverage=1 00:18:31.974 --rc genhtml_legend=1 00:18:31.974 --rc geninfo_all_blocks=1 00:18:31.974 --rc geninfo_unexecuted_blocks=1 00:18:31.974 00:18:31.974 ' 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:18:31.974 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:31.974 --rc genhtml_branch_coverage=1 00:18:31.974 --rc genhtml_function_coverage=1 00:18:31.974 --rc genhtml_legend=1 00:18:31.974 --rc geninfo_all_blocks=1 00:18:31.974 --rc geninfo_unexecuted_blocks=1 00:18:31.974 00:18:31.974 ' 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:18:31.974 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:31.974 --rc genhtml_branch_coverage=1 00:18:31.974 --rc genhtml_function_coverage=1 00:18:31.974 --rc genhtml_legend=1 00:18:31.974 --rc geninfo_all_blocks=1 00:18:31.974 --rc geninfo_unexecuted_blocks=1 00:18:31.974 00:18:31.974 ' 00:18:31.974 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:18:31.975 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:31.975 --rc genhtml_branch_coverage=1 00:18:31.975 --rc genhtml_function_coverage=1 00:18:31.975 --rc genhtml_legend=1 00:18:31.975 --rc geninfo_all_blocks=1 00:18:31.975 --rc geninfo_unexecuted_blocks=1 00:18:31.975 00:18:31.975 ' 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # uname -s 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@15 -- # shopt -s extglob 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@5 -- # export PATH 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@51 -- # : 0 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:18:31.975 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@55 -- # have_pci_nics=0 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@63 -- # nvmftestinit 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@474 -- # prepare_net_devs 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@436 -- # local -g is_hw=no 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@438 -- # remove_spdk_ns 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@309 -- # xtrace_disable 00:18:31.975 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # pci_devs=() 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # local -a pci_devs 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # pci_net_devs=() 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # pci_drivers=() 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # local -A pci_drivers 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # net_devs=() 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # local -ga net_devs 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # e810=() 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # local -ga e810 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # x722=() 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # local -ga x722 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # mlx=() 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # local -ga mlx 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:18:40.113 Found 0000:31:00.0 (0x8086 - 0x159b) 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:40.113 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:18:40.114 Found 0000:31:00.1 (0x8086 - 0x159b) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ up == up ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:18:40.114 Found net devices under 0000:31:00.0: cvl_0_0 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ up == up ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:18:40.114 Found net devices under 0000:31:00.1: cvl_0_1 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # is_hw=yes 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:18:40.114 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:18:40.114 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.711 ms 00:18:40.114 00:18:40.114 --- 10.0.0.2 ping statistics --- 00:18:40.114 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:40.114 rtt min/avg/max/mdev = 0.711/0.711/0.711/0.000 ms 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:18:40.114 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:18:40.114 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.326 ms 00:18:40.114 00:18:40.114 --- 10.0.0.1 ping statistics --- 00:18:40.114 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:40.114 rtt min/avg/max/mdev = 0.326/0.326/0.326/0.000 ms 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@448 -- # return 0 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@64 -- # nvmfappstart -m 0x2 --wait-for-rpc 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=58722 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 58722 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 --wait-for-rpc 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 58722 ']' 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:40.114 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:40.114 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:40.114 [2024-10-17 17:39:47.567344] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:18:40.114 [2024-10-17 17:39:47.567414] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:40.114 [2024-10-17 17:39:47.647741] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:40.114 [2024-10-17 17:39:47.699168] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:40.114 [2024-10-17 17:39:47.699224] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:40.114 [2024-10-17 17:39:47.699233] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:40.114 [2024-10-17 17:39:47.699241] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:40.114 [2024-10-17 17:39:47.699247] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:40.114 [2024-10-17 17:39:47.700070] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@66 -- # '[' tcp '!=' tcp ']' 00:18:40.686 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_set_default_impl -i ssl 00:18:40.947 true 00:18:40.947 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:40.947 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # jq -r .tls_version 00:18:40.948 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # version=0 00:18:40.948 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@75 -- # [[ 0 != \0 ]] 00:18:40.948 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:18:41.208 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:41.208 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # jq -r .tls_version 00:18:41.469 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # version=13 00:18:41.469 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@83 -- # [[ 13 != \1\3 ]] 00:18:41.469 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 7 00:18:41.730 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:41.730 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # jq -r .tls_version 00:18:41.730 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # version=7 00:18:41.730 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@91 -- # [[ 7 != \7 ]] 00:18:41.730 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # jq -r .enable_ktls 00:18:41.730 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:41.991 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # ktls=false 00:18:41.991 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@98 -- # [[ false != \f\a\l\s\e ]] 00:18:41.991 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@104 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --enable-ktls 00:18:42.252 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:42.252 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # jq -r .enable_ktls 00:18:42.512 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # ktls=true 00:18:42.512 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@106 -- # [[ true != \t\r\u\e ]] 00:18:42.512 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@112 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --disable-ktls 00:18:42.512 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:42.512 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # jq -r .enable_ktls 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # ktls=false 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@114 -- # [[ false != \f\a\l\s\e ]] 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # format_interchange_psk 00112233445566778899aabbccddeeff 1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@728 -- # local prefix key digest 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # digest=1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@731 -- # python - 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # key=NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # format_interchange_psk ffeeddccbbaa99887766554433221100 1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 ffeeddccbbaa99887766554433221100 1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@728 -- # local prefix key digest 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # key=ffeeddccbbaa99887766554433221100 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # digest=1 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@731 -- # python - 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # key_2=NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # mktemp 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # key_path=/tmp/tmp.rhGBctHwHP 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # mktemp 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # key_2_path=/tmp/tmp.fGOgE4Nk3q 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@125 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@126 -- # echo -n NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@128 -- # chmod 0600 /tmp/tmp.rhGBctHwHP 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@129 -- # chmod 0600 /tmp/tmp.fGOgE4Nk3q 00:18:42.774 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:18:43.035 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@132 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_start_init 00:18:43.296 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@134 -- # setup_nvmf_tgt /tmp/tmp.rhGBctHwHP 00:18:43.296 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.rhGBctHwHP 00:18:43.296 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:18:43.556 [2024-10-17 17:39:51.289332] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:43.556 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:18:43.816 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:18:43.816 [2024-10-17 17:39:51.686265] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:18:43.816 [2024-10-17 17:39:51.686536] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:43.816 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:18:44.076 malloc0 00:18:44.076 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:18:44.336 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.rhGBctHwHP 00:18:44.596 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:18:44.596 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@138 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -S ssl -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 hostnqn:nqn.2016-06.io.spdk:host1' --psk-path /tmp/tmp.rhGBctHwHP 00:18:56.824 Initializing NVMe Controllers 00:18:56.824 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:18:56.824 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:18:56.824 Initialization complete. Launching workers. 00:18:56.824 ======================================================== 00:18:56.824 Latency(us) 00:18:56.824 Device Information : IOPS MiB/s Average min max 00:18:56.824 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 18856.70 73.66 3394.21 1193.86 4013.14 00:18:56.824 ======================================================== 00:18:56.824 Total : 18856.70 73.66 3394.21 1193.86 4013.14 00:18:56.824 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@144 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.rhGBctHwHP 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.rhGBctHwHP 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=61758 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 61758 /var/tmp/bdevperf.sock 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 61758 ']' 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:56.824 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:56.824 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:56.824 [2024-10-17 17:40:02.601232] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:18:56.824 [2024-10-17 17:40:02.601291] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61758 ] 00:18:56.824 [2024-10-17 17:40:02.680191] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:56.825 [2024-10-17 17:40:02.715358] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:18:56.825 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:56.825 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:18:56.825 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.rhGBctHwHP 00:18:56.825 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:18:56.825 [2024-10-17 17:40:03.730940] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:56.825 TLSTESTn1 00:18:56.825 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:18:56.825 Running I/O for 10 seconds... 00:18:58.028 4622.00 IOPS, 18.05 MiB/s [2024-10-17T15:40:07.332Z] 4597.50 IOPS, 17.96 MiB/s [2024-10-17T15:40:08.272Z] 5087.00 IOPS, 19.87 MiB/s [2024-10-17T15:40:09.212Z] 5327.50 IOPS, 20.81 MiB/s [2024-10-17T15:40:10.152Z] 5430.40 IOPS, 21.21 MiB/s [2024-10-17T15:40:11.090Z] 5259.83 IOPS, 20.55 MiB/s [2024-10-17T15:40:12.032Z] 5358.43 IOPS, 20.93 MiB/s [2024-10-17T15:40:12.972Z] 5399.00 IOPS, 21.09 MiB/s [2024-10-17T15:40:14.356Z] 5488.00 IOPS, 21.44 MiB/s [2024-10-17T15:40:14.356Z] 5455.60 IOPS, 21.31 MiB/s 00:19:06.437 Latency(us) 00:19:06.437 [2024-10-17T15:40:14.356Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:06.437 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:06.437 Verification LBA range: start 0x0 length 0x2000 00:19:06.437 TLSTESTn1 : 10.02 5458.74 21.32 0.00 0.00 23410.44 5679.79 32986.45 00:19:06.437 [2024-10-17T15:40:14.356Z] =================================================================================================================== 00:19:06.437 [2024-10-17T15:40:14.356Z] Total : 5458.74 21.32 0.00 0.00 23410.44 5679.79 32986.45 00:19:06.437 { 00:19:06.437 "results": [ 00:19:06.437 { 00:19:06.437 "job": "TLSTESTn1", 00:19:06.437 "core_mask": "0x4", 00:19:06.437 "workload": "verify", 00:19:06.437 "status": "finished", 00:19:06.437 "verify_range": { 00:19:06.437 "start": 0, 00:19:06.437 "length": 8192 00:19:06.437 }, 00:19:06.437 "queue_depth": 128, 00:19:06.437 "io_size": 4096, 00:19:06.437 "runtime": 10.017691, 00:19:06.437 "iops": 5458.7429378686165, 00:19:06.437 "mibps": 21.323214601049283, 00:19:06.437 "io_failed": 0, 00:19:06.437 "io_timeout": 0, 00:19:06.437 "avg_latency_us": 23410.437923585207, 00:19:06.437 "min_latency_us": 5679.786666666667, 00:19:06.437 "max_latency_us": 32986.45333333333 00:19:06.437 } 00:19:06.437 ], 00:19:06.437 "core_count": 1 00:19:06.437 } 00:19:06.437 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:19:06.437 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 61758 00:19:06.437 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 61758 ']' 00:19:06.437 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 61758 00:19:06.437 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:06.437 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:06.437 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 61758 00:19:06.437 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 61758' 00:19:06.438 killing process with pid 61758 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 61758 00:19:06.438 Received shutdown signal, test time was about 10.000000 seconds 00:19:06.438 00:19:06.438 Latency(us) 00:19:06.438 [2024-10-17T15:40:14.357Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:06.438 [2024-10-17T15:40:14.357Z] =================================================================================================================== 00:19:06.438 [2024-10-17T15:40:14.357Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 61758 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@147 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.fGOgE4Nk3q 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.fGOgE4Nk3q 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.fGOgE4Nk3q 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.fGOgE4Nk3q 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=64000 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 64000 /var/tmp/bdevperf.sock 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 64000 ']' 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:06.438 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:06.438 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:06.438 [2024-10-17 17:40:14.204850] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:06.438 [2024-10-17 17:40:14.204903] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid64000 ] 00:19:06.438 [2024-10-17 17:40:14.281400] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:06.438 [2024-10-17 17:40:14.310843] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:07.377 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:07.377 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:07.377 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.fGOgE4Nk3q 00:19:07.377 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:19:07.637 [2024-10-17 17:40:15.345427] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:07.637 [2024-10-17 17:40:15.352412] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:19:07.637 [2024-10-17 17:40:15.353404] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x17168d0 (107): Transport endpoint is not connected 00:19:07.637 [2024-10-17 17:40:15.354400] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x17168d0 (9): Bad file descriptor 00:19:07.637 [2024-10-17 17:40:15.355401] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:19:07.637 [2024-10-17 17:40:15.355409] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:19:07.637 [2024-10-17 17:40:15.355414] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:19:07.638 [2024-10-17 17:40:15.355422] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:19:07.638 request: 00:19:07.638 { 00:19:07.638 "name": "TLSTEST", 00:19:07.638 "trtype": "tcp", 00:19:07.638 "traddr": "10.0.0.2", 00:19:07.638 "adrfam": "ipv4", 00:19:07.638 "trsvcid": "4420", 00:19:07.638 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:07.638 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:07.638 "prchk_reftag": false, 00:19:07.638 "prchk_guard": false, 00:19:07.638 "hdgst": false, 00:19:07.638 "ddgst": false, 00:19:07.638 "psk": "key0", 00:19:07.638 "allow_unrecognized_csi": false, 00:19:07.638 "method": "bdev_nvme_attach_controller", 00:19:07.638 "req_id": 1 00:19:07.638 } 00:19:07.638 Got JSON-RPC error response 00:19:07.638 response: 00:19:07.638 { 00:19:07.638 "code": -5, 00:19:07.638 "message": "Input/output error" 00:19:07.638 } 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 64000 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 64000 ']' 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 64000 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64000 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64000' 00:19:07.638 killing process with pid 64000 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 64000 00:19:07.638 Received shutdown signal, test time was about 10.000000 seconds 00:19:07.638 00:19:07.638 Latency(us) 00:19:07.638 [2024-10-17T15:40:15.557Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:07.638 [2024-10-17T15:40:15.557Z] =================================================================================================================== 00:19:07.638 [2024-10-17T15:40:15.557Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 64000 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@150 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.rhGBctHwHP 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.rhGBctHwHP 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.rhGBctHwHP 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host2 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.rhGBctHwHP 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=64169 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 64169 /var/tmp/bdevperf.sock 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 64169 ']' 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:07.638 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:07.638 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:07.899 [2024-10-17 17:40:15.583006] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:07.899 [2024-10-17 17:40:15.583058] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid64169 ] 00:19:07.899 [2024-10-17 17:40:15.658541] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:07.899 [2024-10-17 17:40:15.687235] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:08.469 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:08.469 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:08.469 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.rhGBctHwHP 00:19:08.729 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 --psk key0 00:19:08.990 [2024-10-17 17:40:16.673653] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:08.990 [2024-10-17 17:40:16.678006] tcp.c: 969:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:19:08.990 [2024-10-17 17:40:16.678028] posix.c: 574:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:19:08.990 [2024-10-17 17:40:16.678048] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:19:08.990 [2024-10-17 17:40:16.678730] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1eae8d0 (107): Transport endpoint is not connected 00:19:08.990 [2024-10-17 17:40:16.679725] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1eae8d0 (9): Bad file descriptor 00:19:08.990 [2024-10-17 17:40:16.680726] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:19:08.990 [2024-10-17 17:40:16.680733] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:19:08.990 [2024-10-17 17:40:16.680739] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:19:08.990 [2024-10-17 17:40:16.680746] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:19:08.990 request: 00:19:08.990 { 00:19:08.990 "name": "TLSTEST", 00:19:08.990 "trtype": "tcp", 00:19:08.990 "traddr": "10.0.0.2", 00:19:08.990 "adrfam": "ipv4", 00:19:08.990 "trsvcid": "4420", 00:19:08.990 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:08.990 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:19:08.990 "prchk_reftag": false, 00:19:08.990 "prchk_guard": false, 00:19:08.990 "hdgst": false, 00:19:08.990 "ddgst": false, 00:19:08.990 "psk": "key0", 00:19:08.990 "allow_unrecognized_csi": false, 00:19:08.990 "method": "bdev_nvme_attach_controller", 00:19:08.990 "req_id": 1 00:19:08.990 } 00:19:08.990 Got JSON-RPC error response 00:19:08.990 response: 00:19:08.990 { 00:19:08.990 "code": -5, 00:19:08.990 "message": "Input/output error" 00:19:08.990 } 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 64169 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 64169 ']' 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 64169 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64169 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64169' 00:19:08.990 killing process with pid 64169 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 64169 00:19:08.990 Received shutdown signal, test time was about 10.000000 seconds 00:19:08.990 00:19:08.990 Latency(us) 00:19:08.990 [2024-10-17T15:40:16.909Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:08.990 [2024-10-17T15:40:16.909Z] =================================================================================================================== 00:19:08.990 [2024-10-17T15:40:16.909Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 64169 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@153 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.rhGBctHwHP 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.rhGBctHwHP 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.rhGBctHwHP 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode2 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.rhGBctHwHP 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=64466 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 64466 /var/tmp/bdevperf.sock 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 64466 ']' 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:08.990 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:08.990 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:09.262 [2024-10-17 17:40:16.924771] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:09.262 [2024-10-17 17:40:16.924825] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid64466 ] 00:19:09.262 [2024-10-17 17:40:17.003120] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:09.262 [2024-10-17 17:40:17.032151] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:09.832 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:09.832 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:09.832 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.rhGBctHwHP 00:19:10.092 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -q nqn.2016-06.io.spdk:host1 --psk key0 00:19:10.352 [2024-10-17 17:40:18.046504] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:10.352 [2024-10-17 17:40:18.050962] tcp.c: 969:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:19:10.352 [2024-10-17 17:40:18.050983] posix.c: 574:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:19:10.352 [2024-10-17 17:40:18.051002] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:19:10.352 [2024-10-17 17:40:18.051655] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x166c8d0 (107): Transport endpoint is not connected 00:19:10.352 [2024-10-17 17:40:18.052651] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x166c8d0 (9): Bad file descriptor 00:19:10.352 [2024-10-17 17:40:18.053652] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2] Ctrlr is in error state 00:19:10.352 [2024-10-17 17:40:18.053661] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:19:10.352 [2024-10-17 17:40:18.053667] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode2, Operation not permitted 00:19:10.352 [2024-10-17 17:40:18.053674] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2] in failed state. 00:19:10.352 request: 00:19:10.352 { 00:19:10.352 "name": "TLSTEST", 00:19:10.352 "trtype": "tcp", 00:19:10.352 "traddr": "10.0.0.2", 00:19:10.352 "adrfam": "ipv4", 00:19:10.352 "trsvcid": "4420", 00:19:10.352 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:19:10.352 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:10.352 "prchk_reftag": false, 00:19:10.352 "prchk_guard": false, 00:19:10.352 "hdgst": false, 00:19:10.352 "ddgst": false, 00:19:10.352 "psk": "key0", 00:19:10.352 "allow_unrecognized_csi": false, 00:19:10.352 "method": "bdev_nvme_attach_controller", 00:19:10.352 "req_id": 1 00:19:10.352 } 00:19:10.352 Got JSON-RPC error response 00:19:10.352 response: 00:19:10.352 { 00:19:10.352 "code": -5, 00:19:10.352 "message": "Input/output error" 00:19:10.352 } 00:19:10.352 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 64466 00:19:10.352 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 64466 ']' 00:19:10.352 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 64466 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64466 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64466' 00:19:10.353 killing process with pid 64466 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 64466 00:19:10.353 Received shutdown signal, test time was about 10.000000 seconds 00:19:10.353 00:19:10.353 Latency(us) 00:19:10.353 [2024-10-17T15:40:18.272Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:10.353 [2024-10-17T15:40:18.272Z] =================================================================================================================== 00:19:10.353 [2024-10-17T15:40:18.272Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 64466 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@156 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk= 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=64811 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 64811 /var/tmp/bdevperf.sock 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 64811 ']' 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:10.353 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:10.353 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:10.613 [2024-10-17 17:40:18.304873] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:10.613 [2024-10-17 17:40:18.304930] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid64811 ] 00:19:10.613 [2024-10-17 17:40:18.382619] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:10.613 [2024-10-17 17:40:18.410549] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:11.183 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:11.183 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:11.183 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 '' 00:19:11.443 [2024-10-17 17:40:19.244507] keyring.c: 24:keyring_file_check_path: *ERROR*: Non-absolute paths are not allowed: 00:19:11.443 [2024-10-17 17:40:19.244534] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:19:11.443 request: 00:19:11.443 { 00:19:11.443 "name": "key0", 00:19:11.443 "path": "", 00:19:11.443 "method": "keyring_file_add_key", 00:19:11.443 "req_id": 1 00:19:11.443 } 00:19:11.443 Got JSON-RPC error response 00:19:11.444 response: 00:19:11.444 { 00:19:11.444 "code": -1, 00:19:11.444 "message": "Operation not permitted" 00:19:11.444 } 00:19:11.444 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:19:11.705 [2024-10-17 17:40:19.425037] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:11.705 [2024-10-17 17:40:19.425065] bdev_nvme.c:6389:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:19:11.705 request: 00:19:11.705 { 00:19:11.705 "name": "TLSTEST", 00:19:11.705 "trtype": "tcp", 00:19:11.705 "traddr": "10.0.0.2", 00:19:11.705 "adrfam": "ipv4", 00:19:11.705 "trsvcid": "4420", 00:19:11.705 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:11.705 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:11.705 "prchk_reftag": false, 00:19:11.705 "prchk_guard": false, 00:19:11.705 "hdgst": false, 00:19:11.705 "ddgst": false, 00:19:11.705 "psk": "key0", 00:19:11.705 "allow_unrecognized_csi": false, 00:19:11.705 "method": "bdev_nvme_attach_controller", 00:19:11.705 "req_id": 1 00:19:11.705 } 00:19:11.705 Got JSON-RPC error response 00:19:11.705 response: 00:19:11.705 { 00:19:11.705 "code": -126, 00:19:11.705 "message": "Required key not available" 00:19:11.705 } 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 64811 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 64811 ']' 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 64811 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64811 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64811' 00:19:11.705 killing process with pid 64811 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 64811 00:19:11.705 Received shutdown signal, test time was about 10.000000 seconds 00:19:11.705 00:19:11.705 Latency(us) 00:19:11.705 [2024-10-17T15:40:19.624Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:11.705 [2024-10-17T15:40:19.624Z] =================================================================================================================== 00:19:11.705 [2024-10-17T15:40:19.624Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 64811 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@159 -- # killprocess 58722 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 58722 ']' 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 58722 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:11.705 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 58722 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 58722' 00:19:11.966 killing process with pid 58722 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 58722 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 58722 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # format_interchange_psk 00112233445566778899aabbccddeeff0011223344556677 2 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff0011223344556677 2 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@728 -- # local prefix key digest 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff0011223344556677 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # digest=2 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@731 -- # python - 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # key_long=NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # mktemp 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # key_long_path=/tmp/tmp.wGvnMMapw5 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@162 -- # echo -n NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@163 -- # chmod 0600 /tmp/tmp.wGvnMMapw5 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@164 -- # nvmfappstart -m 0x2 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=65166 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 65166 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 65166 ']' 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:11.966 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:11.966 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:12.227 [2024-10-17 17:40:19.912008] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:12.227 [2024-10-17 17:40:19.912077] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:12.227 [2024-10-17 17:40:20.000778] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:12.227 [2024-10-17 17:40:20.037626] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:12.227 [2024-10-17 17:40:20.037664] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:12.227 [2024-10-17 17:40:20.037670] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:12.227 [2024-10-17 17:40:20.037675] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:12.227 [2024-10-17 17:40:20.037679] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:12.227 [2024-10-17 17:40:20.038189] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:12.796 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:12.796 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:12.796 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:19:12.796 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:19:12.796 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:13.057 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:13.057 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@166 -- # setup_nvmf_tgt /tmp/tmp.wGvnMMapw5 00:19:13.057 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wGvnMMapw5 00:19:13.057 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:13.057 [2024-10-17 17:40:20.905700] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:13.057 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:13.317 17:40:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:13.577 [2024-10-17 17:40:21.278606] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:13.577 [2024-10-17 17:40:21.278816] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:13.577 17:40:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:13.577 malloc0 00:19:13.837 17:40:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:13.837 17:40:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:14.099 17:40:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@168 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wGvnMMapw5 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.wGvnMMapw5 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=65530 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 65530 /var/tmp/bdevperf.sock 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 65530 ']' 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:14.359 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:14.359 [2024-10-17 17:40:22.072524] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:14.359 [2024-10-17 17:40:22.072578] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65530 ] 00:19:14.359 [2024-10-17 17:40:22.150087] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:14.359 [2024-10-17 17:40:22.178879] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:14.359 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:14.621 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:19:14.882 [2024-10-17 17:40:22.587931] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:14.882 TLSTESTn1 00:19:14.882 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:19:14.882 Running I/O for 10 seconds... 00:19:17.205 6105.00 IOPS, 23.85 MiB/s [2024-10-17T15:40:26.063Z] 5966.00 IOPS, 23.30 MiB/s [2024-10-17T15:40:27.002Z] 5879.67 IOPS, 22.97 MiB/s [2024-10-17T15:40:27.942Z] 5834.50 IOPS, 22.79 MiB/s [2024-10-17T15:40:28.882Z] 5814.80 IOPS, 22.71 MiB/s [2024-10-17T15:40:29.822Z] 5712.83 IOPS, 22.32 MiB/s [2024-10-17T15:40:31.203Z] 5709.57 IOPS, 22.30 MiB/s [2024-10-17T15:40:32.143Z] 5755.62 IOPS, 22.48 MiB/s [2024-10-17T15:40:33.162Z] 5788.22 IOPS, 22.61 MiB/s [2024-10-17T15:40:33.163Z] 5765.10 IOPS, 22.52 MiB/s 00:19:25.244 Latency(us) 00:19:25.244 [2024-10-17T15:40:33.163Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:25.244 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:25.244 Verification LBA range: start 0x0 length 0x2000 00:19:25.244 TLSTESTn1 : 10.01 5769.58 22.54 0.00 0.00 22154.64 4560.21 35826.35 00:19:25.244 [2024-10-17T15:40:33.163Z] =================================================================================================================== 00:19:25.244 [2024-10-17T15:40:33.163Z] Total : 5769.58 22.54 0.00 0.00 22154.64 4560.21 35826.35 00:19:25.244 { 00:19:25.244 "results": [ 00:19:25.244 { 00:19:25.244 "job": "TLSTESTn1", 00:19:25.244 "core_mask": "0x4", 00:19:25.244 "workload": "verify", 00:19:25.244 "status": "finished", 00:19:25.244 "verify_range": { 00:19:25.244 "start": 0, 00:19:25.244 "length": 8192 00:19:25.244 }, 00:19:25.244 "queue_depth": 128, 00:19:25.244 "io_size": 4096, 00:19:25.244 "runtime": 10.014415, 00:19:25.244 "iops": 5769.5831458951925, 00:19:25.244 "mibps": 22.537434163653096, 00:19:25.244 "io_failed": 0, 00:19:25.244 "io_timeout": 0, 00:19:25.244 "avg_latency_us": 22154.643161933112, 00:19:25.244 "min_latency_us": 4560.213333333333, 00:19:25.244 "max_latency_us": 35826.346666666665 00:19:25.244 } 00:19:25.244 ], 00:19:25.244 "core_count": 1 00:19:25.244 } 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 65530 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 65530 ']' 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 65530 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 65530 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 65530' 00:19:25.244 killing process with pid 65530 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 65530 00:19:25.244 Received shutdown signal, test time was about 10.000000 seconds 00:19:25.244 00:19:25.244 Latency(us) 00:19:25.244 [2024-10-17T15:40:33.163Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:25.244 [2024-10-17T15:40:33.163Z] =================================================================================================================== 00:19:25.244 [2024-10-17T15:40:33.163Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 65530 00:19:25.244 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@171 -- # chmod 0666 /tmp/tmp.wGvnMMapw5 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@172 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wGvnMMapw5 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wGvnMMapw5 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.wGvnMMapw5 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.wGvnMMapw5 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=67639 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 67639 /var/tmp/bdevperf.sock 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 67639 ']' 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:25.244 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:25.244 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:25.244 [2024-10-17 17:40:33.058323] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:25.244 [2024-10-17 17:40:33.058383] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67639 ] 00:19:25.244 [2024-10-17 17:40:33.134635] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:25.528 [2024-10-17 17:40:33.163276] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:25.528 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:25.529 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:25.529 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:25.529 [2024-10-17 17:40:33.400107] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.wGvnMMapw5': 0100666 00:19:25.529 [2024-10-17 17:40:33.400135] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:19:25.529 request: 00:19:25.529 { 00:19:25.529 "name": "key0", 00:19:25.529 "path": "/tmp/tmp.wGvnMMapw5", 00:19:25.529 "method": "keyring_file_add_key", 00:19:25.529 "req_id": 1 00:19:25.529 } 00:19:25.529 Got JSON-RPC error response 00:19:25.529 response: 00:19:25.529 { 00:19:25.529 "code": -1, 00:19:25.529 "message": "Operation not permitted" 00:19:25.529 } 00:19:25.529 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:19:25.789 [2024-10-17 17:40:33.588650] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:25.789 [2024-10-17 17:40:33.588685] bdev_nvme.c:6389:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:19:25.789 request: 00:19:25.789 { 00:19:25.789 "name": "TLSTEST", 00:19:25.789 "trtype": "tcp", 00:19:25.789 "traddr": "10.0.0.2", 00:19:25.789 "adrfam": "ipv4", 00:19:25.789 "trsvcid": "4420", 00:19:25.789 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:25.789 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:25.789 "prchk_reftag": false, 00:19:25.789 "prchk_guard": false, 00:19:25.789 "hdgst": false, 00:19:25.789 "ddgst": false, 00:19:25.789 "psk": "key0", 00:19:25.789 "allow_unrecognized_csi": false, 00:19:25.789 "method": "bdev_nvme_attach_controller", 00:19:25.789 "req_id": 1 00:19:25.789 } 00:19:25.789 Got JSON-RPC error response 00:19:25.789 response: 00:19:25.789 { 00:19:25.789 "code": -126, 00:19:25.789 "message": "Required key not available" 00:19:25.789 } 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 67639 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 67639 ']' 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 67639 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 67639 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 67639' 00:19:25.789 killing process with pid 67639 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 67639 00:19:25.789 Received shutdown signal, test time was about 10.000000 seconds 00:19:25.789 00:19:25.789 Latency(us) 00:19:25.789 [2024-10-17T15:40:33.708Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:25.789 [2024-10-17T15:40:33.708Z] =================================================================================================================== 00:19:25.789 [2024-10-17T15:40:33.708Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:25.789 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 67639 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@175 -- # killprocess 65166 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 65166 ']' 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 65166 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 65166 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 65166' 00:19:26.049 killing process with pid 65166 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 65166 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 65166 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@176 -- # nvmfappstart -m 0x2 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=67896 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 67896 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 67896 ']' 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:26.049 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:26.049 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:26.309 [2024-10-17 17:40:33.998587] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:26.309 [2024-10-17 17:40:33.998644] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:26.309 [2024-10-17 17:40:34.083041] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:26.309 [2024-10-17 17:40:34.111912] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:26.309 [2024-10-17 17:40:34.111946] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:26.309 [2024-10-17 17:40:34.111952] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:26.309 [2024-10-17 17:40:34.111957] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:26.309 [2024-10-17 17:40:34.111961] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:26.309 [2024-10-17 17:40:34.112463] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:26.879 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:26.879 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@178 -- # NOT setup_nvmf_tgt /tmp/tmp.wGvnMMapw5 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg setup_nvmf_tgt /tmp/tmp.wGvnMMapw5 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=setup_nvmf_tgt 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t setup_nvmf_tgt 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # setup_nvmf_tgt /tmp/tmp.wGvnMMapw5 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wGvnMMapw5 00:19:27.139 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:27.139 [2024-10-17 17:40:34.997126] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:27.139 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:27.399 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:27.658 [2024-10-17 17:40:35.362024] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:27.658 [2024-10-17 17:40:35.362226] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:27.658 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:27.658 malloc0 00:19:27.918 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:27.918 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:28.178 [2024-10-17 17:40:35.905125] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.wGvnMMapw5': 0100666 00:19:28.178 [2024-10-17 17:40:35.905145] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:19:28.178 request: 00:19:28.178 { 00:19:28.178 "name": "key0", 00:19:28.178 "path": "/tmp/tmp.wGvnMMapw5", 00:19:28.178 "method": "keyring_file_add_key", 00:19:28.178 "req_id": 1 00:19:28.178 } 00:19:28.178 Got JSON-RPC error response 00:19:28.178 response: 00:19:28.178 { 00:19:28.178 "code": -1, 00:19:28.178 "message": "Operation not permitted" 00:19:28.178 } 00:19:28.178 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:19:28.178 [2024-10-17 17:40:36.081585] tcp.c:3792:nvmf_tcp_subsystem_add_host: *ERROR*: Key 'key0' does not exist 00:19:28.178 [2024-10-17 17:40:36.081612] subsystem.c:1055:spdk_nvmf_subsystem_add_host_ext: *ERROR*: Unable to add host to TCP transport 00:19:28.178 request: 00:19:28.178 { 00:19:28.178 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.178 "host": "nqn.2016-06.io.spdk:host1", 00:19:28.178 "psk": "key0", 00:19:28.178 "method": "nvmf_subsystem_add_host", 00:19:28.178 "req_id": 1 00:19:28.178 } 00:19:28.178 Got JSON-RPC error response 00:19:28.178 response: 00:19:28.178 { 00:19:28.178 "code": -32603, 00:19:28.178 "message": "Internal error" 00:19:28.178 } 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@181 -- # killprocess 67896 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 67896 ']' 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 67896 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 67896 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 67896' 00:19:28.438 killing process with pid 67896 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 67896 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 67896 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@182 -- # chmod 0600 /tmp/tmp.wGvnMMapw5 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@185 -- # nvmfappstart -m 0x2 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=68286 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 68286 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 68286 ']' 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:28.438 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:28.438 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:28.438 [2024-10-17 17:40:36.353305] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:28.438 [2024-10-17 17:40:36.353367] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:28.699 [2024-10-17 17:40:36.435913] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:28.699 [2024-10-17 17:40:36.468357] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:28.699 [2024-10-17 17:40:36.468388] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:28.699 [2024-10-17 17:40:36.468394] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:28.699 [2024-10-17 17:40:36.468399] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:28.699 [2024-10-17 17:40:36.468403] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:28.699 [2024-10-17 17:40:36.468930] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@186 -- # setup_nvmf_tgt /tmp/tmp.wGvnMMapw5 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wGvnMMapw5 00:19:29.270 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:29.529 [2024-10-17 17:40:37.342367] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:29.529 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:29.789 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:29.789 [2024-10-17 17:40:37.703252] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:29.789 [2024-10-17 17:40:37.703458] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:30.050 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:30.050 malloc0 00:19:30.050 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:30.310 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@189 -- # bdevperf_pid=68836 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@188 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@191 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@192 -- # waitforlisten 68836 /var/tmp/bdevperf.sock 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 68836 ']' 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:30.571 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:30.571 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:30.832 [2024-10-17 17:40:38.500174] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:30.832 [2024-10-17 17:40:38.500232] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68836 ] 00:19:30.832 [2024-10-17 17:40:38.575927] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:30.832 [2024-10-17 17:40:38.611265] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:31.411 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:31.411 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:31.411 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@193 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:31.671 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@194 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:19:31.931 [2024-10-17 17:40:39.618722] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:31.931 TLSTESTn1 00:19:31.931 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py save_config 00:19:32.193 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # tgtconf='{ 00:19:32.193 "subsystems": [ 00:19:32.193 { 00:19:32.193 "subsystem": "keyring", 00:19:32.193 "config": [ 00:19:32.193 { 00:19:32.193 "method": "keyring_file_add_key", 00:19:32.193 "params": { 00:19:32.193 "name": "key0", 00:19:32.193 "path": "/tmp/tmp.wGvnMMapw5" 00:19:32.193 } 00:19:32.193 } 00:19:32.193 ] 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "subsystem": "iobuf", 00:19:32.193 "config": [ 00:19:32.193 { 00:19:32.193 "method": "iobuf_set_options", 00:19:32.193 "params": { 00:19:32.193 "small_pool_count": 8192, 00:19:32.193 "large_pool_count": 1024, 00:19:32.193 "small_bufsize": 8192, 00:19:32.193 "large_bufsize": 135168 00:19:32.193 } 00:19:32.193 } 00:19:32.193 ] 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "subsystem": "sock", 00:19:32.193 "config": [ 00:19:32.193 { 00:19:32.193 "method": "sock_set_default_impl", 00:19:32.193 "params": { 00:19:32.193 "impl_name": "posix" 00:19:32.193 } 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "method": "sock_impl_set_options", 00:19:32.193 "params": { 00:19:32.193 "impl_name": "ssl", 00:19:32.193 "recv_buf_size": 4096, 00:19:32.193 "send_buf_size": 4096, 00:19:32.193 "enable_recv_pipe": true, 00:19:32.193 "enable_quickack": false, 00:19:32.193 "enable_placement_id": 0, 00:19:32.193 "enable_zerocopy_send_server": true, 00:19:32.193 "enable_zerocopy_send_client": false, 00:19:32.193 "zerocopy_threshold": 0, 00:19:32.193 "tls_version": 0, 00:19:32.193 "enable_ktls": false 00:19:32.193 } 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "method": "sock_impl_set_options", 00:19:32.193 "params": { 00:19:32.193 "impl_name": "posix", 00:19:32.193 "recv_buf_size": 2097152, 00:19:32.193 "send_buf_size": 2097152, 00:19:32.193 "enable_recv_pipe": true, 00:19:32.193 "enable_quickack": false, 00:19:32.193 "enable_placement_id": 0, 00:19:32.193 "enable_zerocopy_send_server": true, 00:19:32.193 "enable_zerocopy_send_client": false, 00:19:32.193 "zerocopy_threshold": 0, 00:19:32.193 "tls_version": 0, 00:19:32.193 "enable_ktls": false 00:19:32.193 } 00:19:32.193 } 00:19:32.193 ] 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "subsystem": "vmd", 00:19:32.193 "config": [] 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "subsystem": "accel", 00:19:32.193 "config": [ 00:19:32.193 { 00:19:32.193 "method": "accel_set_options", 00:19:32.193 "params": { 00:19:32.193 "small_cache_size": 128, 00:19:32.193 "large_cache_size": 16, 00:19:32.193 "task_count": 2048, 00:19:32.193 "sequence_count": 2048, 00:19:32.193 "buf_count": 2048 00:19:32.193 } 00:19:32.193 } 00:19:32.193 ] 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "subsystem": "bdev", 00:19:32.193 "config": [ 00:19:32.193 { 00:19:32.193 "method": "bdev_set_options", 00:19:32.193 "params": { 00:19:32.193 "bdev_io_pool_size": 65535, 00:19:32.193 "bdev_io_cache_size": 256, 00:19:32.193 "bdev_auto_examine": true, 00:19:32.193 "iobuf_small_cache_size": 128, 00:19:32.193 "iobuf_large_cache_size": 16 00:19:32.193 } 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "method": "bdev_raid_set_options", 00:19:32.193 "params": { 00:19:32.193 "process_window_size_kb": 1024, 00:19:32.193 "process_max_bandwidth_mb_sec": 0 00:19:32.193 } 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "method": "bdev_iscsi_set_options", 00:19:32.193 "params": { 00:19:32.193 "timeout_sec": 30 00:19:32.193 } 00:19:32.193 }, 00:19:32.193 { 00:19:32.193 "method": "bdev_nvme_set_options", 00:19:32.193 "params": { 00:19:32.193 "action_on_timeout": "none", 00:19:32.193 "timeout_us": 0, 00:19:32.193 "timeout_admin_us": 0, 00:19:32.193 "keep_alive_timeout_ms": 10000, 00:19:32.193 "arbitration_burst": 0, 00:19:32.193 "low_priority_weight": 0, 00:19:32.193 "medium_priority_weight": 0, 00:19:32.193 "high_priority_weight": 0, 00:19:32.193 "nvme_adminq_poll_period_us": 10000, 00:19:32.193 "nvme_ioq_poll_period_us": 0, 00:19:32.193 "io_queue_requests": 0, 00:19:32.193 "delay_cmd_submit": true, 00:19:32.193 "transport_retry_count": 4, 00:19:32.193 "bdev_retry_count": 3, 00:19:32.193 "transport_ack_timeout": 0, 00:19:32.193 "ctrlr_loss_timeout_sec": 0, 00:19:32.193 "reconnect_delay_sec": 0, 00:19:32.193 "fast_io_fail_timeout_sec": 0, 00:19:32.193 "disable_auto_failback": false, 00:19:32.193 "generate_uuids": false, 00:19:32.193 "transport_tos": 0, 00:19:32.193 "nvme_error_stat": false, 00:19:32.193 "rdma_srq_size": 0, 00:19:32.193 "io_path_stat": false, 00:19:32.193 "allow_accel_sequence": false, 00:19:32.193 "rdma_max_cq_size": 0, 00:19:32.193 "rdma_cm_event_timeout_ms": 0, 00:19:32.193 "dhchap_digests": [ 00:19:32.193 "sha256", 00:19:32.193 "sha384", 00:19:32.193 "sha512" 00:19:32.193 ], 00:19:32.193 "dhchap_dhgroups": [ 00:19:32.193 "null", 00:19:32.193 "ffdhe2048", 00:19:32.193 "ffdhe3072", 00:19:32.193 "ffdhe4096", 00:19:32.193 "ffdhe6144", 00:19:32.193 "ffdhe8192" 00:19:32.193 ], 00:19:32.194 "rdma_umr_per_io": false 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "bdev_nvme_set_hotplug", 00:19:32.194 "params": { 00:19:32.194 "period_us": 100000, 00:19:32.194 "enable": false 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "bdev_malloc_create", 00:19:32.194 "params": { 00:19:32.194 "name": "malloc0", 00:19:32.194 "num_blocks": 8192, 00:19:32.194 "block_size": 4096, 00:19:32.194 "physical_block_size": 4096, 00:19:32.194 "uuid": "89c81361-2d98-4633-8857-f7878c154989", 00:19:32.194 "optimal_io_boundary": 0, 00:19:32.194 "md_size": 0, 00:19:32.194 "dif_type": 0, 00:19:32.194 "dif_is_head_of_md": false, 00:19:32.194 "dif_pi_format": 0 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "bdev_wait_for_examine" 00:19:32.194 } 00:19:32.194 ] 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "subsystem": "nbd", 00:19:32.194 "config": [] 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "subsystem": "scheduler", 00:19:32.194 "config": [ 00:19:32.194 { 00:19:32.194 "method": "framework_set_scheduler", 00:19:32.194 "params": { 00:19:32.194 "name": "static" 00:19:32.194 } 00:19:32.194 } 00:19:32.194 ] 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "subsystem": "nvmf", 00:19:32.194 "config": [ 00:19:32.194 { 00:19:32.194 "method": "nvmf_set_config", 00:19:32.194 "params": { 00:19:32.194 "discovery_filter": "match_any", 00:19:32.194 "admin_cmd_passthru": { 00:19:32.194 "identify_ctrlr": false 00:19:32.194 }, 00:19:32.194 "dhchap_digests": [ 00:19:32.194 "sha256", 00:19:32.194 "sha384", 00:19:32.194 "sha512" 00:19:32.194 ], 00:19:32.194 "dhchap_dhgroups": [ 00:19:32.194 "null", 00:19:32.194 "ffdhe2048", 00:19:32.194 "ffdhe3072", 00:19:32.194 "ffdhe4096", 00:19:32.194 "ffdhe6144", 00:19:32.194 "ffdhe8192" 00:19:32.194 ] 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "nvmf_set_max_subsystems", 00:19:32.194 "params": { 00:19:32.194 "max_subsystems": 1024 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "nvmf_set_crdt", 00:19:32.194 "params": { 00:19:32.194 "crdt1": 0, 00:19:32.194 "crdt2": 0, 00:19:32.194 "crdt3": 0 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "nvmf_create_transport", 00:19:32.194 "params": { 00:19:32.194 "trtype": "TCP", 00:19:32.194 "max_queue_depth": 128, 00:19:32.194 "max_io_qpairs_per_ctrlr": 127, 00:19:32.194 "in_capsule_data_size": 4096, 00:19:32.194 "max_io_size": 131072, 00:19:32.194 "io_unit_size": 131072, 00:19:32.194 "max_aq_depth": 128, 00:19:32.194 "num_shared_buffers": 511, 00:19:32.194 "buf_cache_size": 4294967295, 00:19:32.194 "dif_insert_or_strip": false, 00:19:32.194 "zcopy": false, 00:19:32.194 "c2h_success": false, 00:19:32.194 "sock_priority": 0, 00:19:32.194 "abort_timeout_sec": 1, 00:19:32.194 "ack_timeout": 0, 00:19:32.194 "data_wr_pool_size": 0 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "nvmf_create_subsystem", 00:19:32.194 "params": { 00:19:32.194 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.194 "allow_any_host": false, 00:19:32.194 "serial_number": "SPDK00000000000001", 00:19:32.194 "model_number": "SPDK bdev Controller", 00:19:32.194 "max_namespaces": 10, 00:19:32.194 "min_cntlid": 1, 00:19:32.194 "max_cntlid": 65519, 00:19:32.194 "ana_reporting": false 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "nvmf_subsystem_add_host", 00:19:32.194 "params": { 00:19:32.194 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.194 "host": "nqn.2016-06.io.spdk:host1", 00:19:32.194 "psk": "key0" 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "nvmf_subsystem_add_ns", 00:19:32.194 "params": { 00:19:32.194 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.194 "namespace": { 00:19:32.194 "nsid": 1, 00:19:32.194 "bdev_name": "malloc0", 00:19:32.194 "nguid": "89C813612D9846338857F7878C154989", 00:19:32.194 "uuid": "89c81361-2d98-4633-8857-f7878c154989", 00:19:32.194 "no_auto_visible": false 00:19:32.194 } 00:19:32.194 } 00:19:32.194 }, 00:19:32.194 { 00:19:32.194 "method": "nvmf_subsystem_add_listener", 00:19:32.194 "params": { 00:19:32.194 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.194 "listen_address": { 00:19:32.194 "trtype": "TCP", 00:19:32.194 "adrfam": "IPv4", 00:19:32.194 "traddr": "10.0.0.2", 00:19:32.194 "trsvcid": "4420" 00:19:32.194 }, 00:19:32.194 "secure_channel": true 00:19:32.194 } 00:19:32.194 } 00:19:32.194 ] 00:19:32.194 } 00:19:32.194 ] 00:19:32.194 }' 00:19:32.194 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:19:32.455 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # bdevperfconf='{ 00:19:32.455 "subsystems": [ 00:19:32.455 { 00:19:32.455 "subsystem": "keyring", 00:19:32.455 "config": [ 00:19:32.455 { 00:19:32.455 "method": "keyring_file_add_key", 00:19:32.455 "params": { 00:19:32.455 "name": "key0", 00:19:32.455 "path": "/tmp/tmp.wGvnMMapw5" 00:19:32.455 } 00:19:32.455 } 00:19:32.455 ] 00:19:32.455 }, 00:19:32.455 { 00:19:32.455 "subsystem": "iobuf", 00:19:32.455 "config": [ 00:19:32.455 { 00:19:32.455 "method": "iobuf_set_options", 00:19:32.455 "params": { 00:19:32.455 "small_pool_count": 8192, 00:19:32.455 "large_pool_count": 1024, 00:19:32.455 "small_bufsize": 8192, 00:19:32.455 "large_bufsize": 135168 00:19:32.455 } 00:19:32.455 } 00:19:32.455 ] 00:19:32.455 }, 00:19:32.455 { 00:19:32.455 "subsystem": "sock", 00:19:32.455 "config": [ 00:19:32.455 { 00:19:32.455 "method": "sock_set_default_impl", 00:19:32.455 "params": { 00:19:32.455 "impl_name": "posix" 00:19:32.455 } 00:19:32.455 }, 00:19:32.455 { 00:19:32.455 "method": "sock_impl_set_options", 00:19:32.455 "params": { 00:19:32.455 "impl_name": "ssl", 00:19:32.455 "recv_buf_size": 4096, 00:19:32.455 "send_buf_size": 4096, 00:19:32.455 "enable_recv_pipe": true, 00:19:32.455 "enable_quickack": false, 00:19:32.455 "enable_placement_id": 0, 00:19:32.455 "enable_zerocopy_send_server": true, 00:19:32.455 "enable_zerocopy_send_client": false, 00:19:32.455 "zerocopy_threshold": 0, 00:19:32.455 "tls_version": 0, 00:19:32.455 "enable_ktls": false 00:19:32.455 } 00:19:32.455 }, 00:19:32.455 { 00:19:32.455 "method": "sock_impl_set_options", 00:19:32.455 "params": { 00:19:32.455 "impl_name": "posix", 00:19:32.455 "recv_buf_size": 2097152, 00:19:32.455 "send_buf_size": 2097152, 00:19:32.455 "enable_recv_pipe": true, 00:19:32.455 "enable_quickack": false, 00:19:32.455 "enable_placement_id": 0, 00:19:32.455 "enable_zerocopy_send_server": true, 00:19:32.455 "enable_zerocopy_send_client": false, 00:19:32.455 "zerocopy_threshold": 0, 00:19:32.455 "tls_version": 0, 00:19:32.455 "enable_ktls": false 00:19:32.455 } 00:19:32.455 } 00:19:32.455 ] 00:19:32.455 }, 00:19:32.455 { 00:19:32.455 "subsystem": "vmd", 00:19:32.455 "config": [] 00:19:32.455 }, 00:19:32.455 { 00:19:32.455 "subsystem": "accel", 00:19:32.455 "config": [ 00:19:32.455 { 00:19:32.455 "method": "accel_set_options", 00:19:32.455 "params": { 00:19:32.455 "small_cache_size": 128, 00:19:32.455 "large_cache_size": 16, 00:19:32.455 "task_count": 2048, 00:19:32.455 "sequence_count": 2048, 00:19:32.455 "buf_count": 2048 00:19:32.455 } 00:19:32.455 } 00:19:32.455 ] 00:19:32.455 }, 00:19:32.455 { 00:19:32.455 "subsystem": "bdev", 00:19:32.455 "config": [ 00:19:32.455 { 00:19:32.455 "method": "bdev_set_options", 00:19:32.455 "params": { 00:19:32.455 "bdev_io_pool_size": 65535, 00:19:32.455 "bdev_io_cache_size": 256, 00:19:32.455 "bdev_auto_examine": true, 00:19:32.455 "iobuf_small_cache_size": 128, 00:19:32.456 "iobuf_large_cache_size": 16 00:19:32.456 } 00:19:32.456 }, 00:19:32.456 { 00:19:32.456 "method": "bdev_raid_set_options", 00:19:32.456 "params": { 00:19:32.456 "process_window_size_kb": 1024, 00:19:32.456 "process_max_bandwidth_mb_sec": 0 00:19:32.456 } 00:19:32.456 }, 00:19:32.456 { 00:19:32.456 "method": "bdev_iscsi_set_options", 00:19:32.456 "params": { 00:19:32.456 "timeout_sec": 30 00:19:32.456 } 00:19:32.456 }, 00:19:32.456 { 00:19:32.456 "method": "bdev_nvme_set_options", 00:19:32.456 "params": { 00:19:32.456 "action_on_timeout": "none", 00:19:32.456 "timeout_us": 0, 00:19:32.456 "timeout_admin_us": 0, 00:19:32.456 "keep_alive_timeout_ms": 10000, 00:19:32.456 "arbitration_burst": 0, 00:19:32.456 "low_priority_weight": 0, 00:19:32.456 "medium_priority_weight": 0, 00:19:32.456 "high_priority_weight": 0, 00:19:32.456 "nvme_adminq_poll_period_us": 10000, 00:19:32.456 "nvme_ioq_poll_period_us": 0, 00:19:32.456 "io_queue_requests": 512, 00:19:32.456 "delay_cmd_submit": true, 00:19:32.456 "transport_retry_count": 4, 00:19:32.456 "bdev_retry_count": 3, 00:19:32.456 "transport_ack_timeout": 0, 00:19:32.456 "ctrlr_loss_timeout_sec": 0, 00:19:32.456 "reconnect_delay_sec": 0, 00:19:32.456 "fast_io_fail_timeout_sec": 0, 00:19:32.456 "disable_auto_failback": false, 00:19:32.456 "generate_uuids": false, 00:19:32.456 "transport_tos": 0, 00:19:32.456 "nvme_error_stat": false, 00:19:32.456 "rdma_srq_size": 0, 00:19:32.456 "io_path_stat": false, 00:19:32.456 "allow_accel_sequence": false, 00:19:32.456 "rdma_max_cq_size": 0, 00:19:32.456 "rdma_cm_event_timeout_ms": 0, 00:19:32.456 "dhchap_digests": [ 00:19:32.456 "sha256", 00:19:32.456 "sha384", 00:19:32.456 "sha512" 00:19:32.456 ], 00:19:32.456 "dhchap_dhgroups": [ 00:19:32.456 "null", 00:19:32.456 "ffdhe2048", 00:19:32.456 "ffdhe3072", 00:19:32.456 "ffdhe4096", 00:19:32.456 "ffdhe6144", 00:19:32.456 "ffdhe8192" 00:19:32.456 ], 00:19:32.456 "rdma_umr_per_io": false 00:19:32.456 } 00:19:32.456 }, 00:19:32.456 { 00:19:32.456 "method": "bdev_nvme_attach_controller", 00:19:32.456 "params": { 00:19:32.456 "name": "TLSTEST", 00:19:32.456 "trtype": "TCP", 00:19:32.456 "adrfam": "IPv4", 00:19:32.456 "traddr": "10.0.0.2", 00:19:32.456 "trsvcid": "4420", 00:19:32.456 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.456 "prchk_reftag": false, 00:19:32.456 "prchk_guard": false, 00:19:32.456 "ctrlr_loss_timeout_sec": 0, 00:19:32.456 "reconnect_delay_sec": 0, 00:19:32.456 "fast_io_fail_timeout_sec": 0, 00:19:32.456 "psk": "key0", 00:19:32.456 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:32.456 "hdgst": false, 00:19:32.456 "ddgst": false, 00:19:32.456 "multipath": "multipath" 00:19:32.456 } 00:19:32.456 }, 00:19:32.456 { 00:19:32.456 "method": "bdev_nvme_set_hotplug", 00:19:32.456 "params": { 00:19:32.456 "period_us": 100000, 00:19:32.456 "enable": false 00:19:32.456 } 00:19:32.456 }, 00:19:32.456 { 00:19:32.456 "method": "bdev_wait_for_examine" 00:19:32.456 } 00:19:32.456 ] 00:19:32.456 }, 00:19:32.456 { 00:19:32.456 "subsystem": "nbd", 00:19:32.456 "config": [] 00:19:32.456 } 00:19:32.456 ] 00:19:32.456 }' 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@201 -- # killprocess 68836 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 68836 ']' 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 68836 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 68836 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 68836' 00:19:32.456 killing process with pid 68836 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 68836 00:19:32.456 Received shutdown signal, test time was about 10.000000 seconds 00:19:32.456 00:19:32.456 Latency(us) 00:19:32.456 [2024-10-17T15:40:40.375Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:32.456 [2024-10-17T15:40:40.375Z] =================================================================================================================== 00:19:32.456 [2024-10-17T15:40:40.375Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:32.456 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 68836 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@202 -- # killprocess 68286 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 68286 ']' 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 68286 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 68286 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 68286' 00:19:32.716 killing process with pid 68286 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 68286 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 68286 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # nvmfappstart -m 0x2 -c /dev/fd/62 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:32.716 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # echo '{ 00:19:32.716 "subsystems": [ 00:19:32.716 { 00:19:32.716 "subsystem": "keyring", 00:19:32.716 "config": [ 00:19:32.716 { 00:19:32.716 "method": "keyring_file_add_key", 00:19:32.716 "params": { 00:19:32.716 "name": "key0", 00:19:32.716 "path": "/tmp/tmp.wGvnMMapw5" 00:19:32.716 } 00:19:32.716 } 00:19:32.716 ] 00:19:32.716 }, 00:19:32.716 { 00:19:32.716 "subsystem": "iobuf", 00:19:32.716 "config": [ 00:19:32.716 { 00:19:32.716 "method": "iobuf_set_options", 00:19:32.716 "params": { 00:19:32.716 "small_pool_count": 8192, 00:19:32.716 "large_pool_count": 1024, 00:19:32.716 "small_bufsize": 8192, 00:19:32.716 "large_bufsize": 135168 00:19:32.716 } 00:19:32.716 } 00:19:32.716 ] 00:19:32.716 }, 00:19:32.716 { 00:19:32.716 "subsystem": "sock", 00:19:32.716 "config": [ 00:19:32.717 { 00:19:32.717 "method": "sock_set_default_impl", 00:19:32.717 "params": { 00:19:32.717 "impl_name": "posix" 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "sock_impl_set_options", 00:19:32.717 "params": { 00:19:32.717 "impl_name": "ssl", 00:19:32.717 "recv_buf_size": 4096, 00:19:32.717 "send_buf_size": 4096, 00:19:32.717 "enable_recv_pipe": true, 00:19:32.717 "enable_quickack": false, 00:19:32.717 "enable_placement_id": 0, 00:19:32.717 "enable_zerocopy_send_server": true, 00:19:32.717 "enable_zerocopy_send_client": false, 00:19:32.717 "zerocopy_threshold": 0, 00:19:32.717 "tls_version": 0, 00:19:32.717 "enable_ktls": false 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "sock_impl_set_options", 00:19:32.717 "params": { 00:19:32.717 "impl_name": "posix", 00:19:32.717 "recv_buf_size": 2097152, 00:19:32.717 "send_buf_size": 2097152, 00:19:32.717 "enable_recv_pipe": true, 00:19:32.717 "enable_quickack": false, 00:19:32.717 "enable_placement_id": 0, 00:19:32.717 "enable_zerocopy_send_server": true, 00:19:32.717 "enable_zerocopy_send_client": false, 00:19:32.717 "zerocopy_threshold": 0, 00:19:32.717 "tls_version": 0, 00:19:32.717 "enable_ktls": false 00:19:32.717 } 00:19:32.717 } 00:19:32.717 ] 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "subsystem": "vmd", 00:19:32.717 "config": [] 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "subsystem": "accel", 00:19:32.717 "config": [ 00:19:32.717 { 00:19:32.717 "method": "accel_set_options", 00:19:32.717 "params": { 00:19:32.717 "small_cache_size": 128, 00:19:32.717 "large_cache_size": 16, 00:19:32.717 "task_count": 2048, 00:19:32.717 "sequence_count": 2048, 00:19:32.717 "buf_count": 2048 00:19:32.717 } 00:19:32.717 } 00:19:32.717 ] 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "subsystem": "bdev", 00:19:32.717 "config": [ 00:19:32.717 { 00:19:32.717 "method": "bdev_set_options", 00:19:32.717 "params": { 00:19:32.717 "bdev_io_pool_size": 65535, 00:19:32.717 "bdev_io_cache_size": 256, 00:19:32.717 "bdev_auto_examine": true, 00:19:32.717 "iobuf_small_cache_size": 128, 00:19:32.717 "iobuf_large_cache_size": 16 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "bdev_raid_set_options", 00:19:32.717 "params": { 00:19:32.717 "process_window_size_kb": 1024, 00:19:32.717 "process_max_bandwidth_mb_sec": 0 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "bdev_iscsi_set_options", 00:19:32.717 "params": { 00:19:32.717 "timeout_sec": 30 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "bdev_nvme_set_options", 00:19:32.717 "params": { 00:19:32.717 "action_on_timeout": "none", 00:19:32.717 "timeout_us": 0, 00:19:32.717 "timeout_admin_us": 0, 00:19:32.717 "keep_alive_timeout_ms": 10000, 00:19:32.717 "arbitration_burst": 0, 00:19:32.717 "low_priority_weight": 0, 00:19:32.717 "medium_priority_weight": 0, 00:19:32.717 "high_priority_weight": 0, 00:19:32.717 "nvme_adminq_poll_period_us": 10000, 00:19:32.717 "nvme_ioq_poll_period_us": 0, 00:19:32.717 "io_queue_requests": 0, 00:19:32.717 "delay_cmd_submit": true, 00:19:32.717 "transport_retry_count": 4, 00:19:32.717 "bdev_retry_count": 3, 00:19:32.717 "transport_ack_timeout": 0, 00:19:32.717 "ctrlr_loss_timeout_sec": 0, 00:19:32.717 "reconnect_delay_sec": 0, 00:19:32.717 "fast_io_fail_timeout_sec": 0, 00:19:32.717 "disable_auto_failback": false, 00:19:32.717 "generate_uuids": false, 00:19:32.717 "transport_tos": 0, 00:19:32.717 "nvme_error_stat": false, 00:19:32.717 "rdma_srq_size": 0, 00:19:32.717 "io_path_stat": false, 00:19:32.717 "allow_accel_sequence": false, 00:19:32.717 "rdma_max_cq_size": 0, 00:19:32.717 "rdma_cm_event_timeout_ms": 0, 00:19:32.717 "dhchap_digests": [ 00:19:32.717 "sha256", 00:19:32.717 "sha384", 00:19:32.717 "sha512" 00:19:32.717 ], 00:19:32.717 "dhchap_dhgroups": [ 00:19:32.717 "null", 00:19:32.717 "ffdhe2048", 00:19:32.717 "ffdhe3072", 00:19:32.717 "ffdhe4096", 00:19:32.717 "ffdhe6144", 00:19:32.717 "ffdhe8192" 00:19:32.717 ], 00:19:32.717 "rdma_umr_per_io": false 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "bdev_nvme_set_hotplug", 00:19:32.717 "params": { 00:19:32.717 "period_us": 100000, 00:19:32.717 "enable": false 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "bdev_malloc_create", 00:19:32.717 "params": { 00:19:32.717 "name": "malloc0", 00:19:32.717 "num_blocks": 8192, 00:19:32.717 "block_size": 4096, 00:19:32.717 "physical_block_size": 4096, 00:19:32.717 "uuid": "89c81361-2d98-4633-8857-f7878c154989", 00:19:32.717 "optimal_io_boundary": 0, 00:19:32.717 "md_size": 0, 00:19:32.717 "dif_type": 0, 00:19:32.717 "dif_is_head_of_md": false, 00:19:32.717 "dif_pi_format": 0 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "bdev_wait_for_examine" 00:19:32.717 } 00:19:32.717 ] 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "subsystem": "nbd", 00:19:32.717 "config": [] 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "subsystem": "scheduler", 00:19:32.717 "config": [ 00:19:32.717 { 00:19:32.717 "method": "framework_set_scheduler", 00:19:32.717 "params": { 00:19:32.717 "name": "static" 00:19:32.717 } 00:19:32.717 } 00:19:32.717 ] 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "subsystem": "nvmf", 00:19:32.717 "config": [ 00:19:32.717 { 00:19:32.717 "method": "nvmf_set_config", 00:19:32.717 "params": { 00:19:32.717 "discovery_filter": "match_any", 00:19:32.717 "admin_cmd_passthru": { 00:19:32.717 "identify_ctrlr": false 00:19:32.717 }, 00:19:32.717 "dhchap_digests": [ 00:19:32.717 "sha256", 00:19:32.717 "sha384", 00:19:32.717 "sha512" 00:19:32.717 ], 00:19:32.717 "dhchap_dhgroups": [ 00:19:32.717 "null", 00:19:32.717 "ffdhe2048", 00:19:32.717 "ffdhe3072", 00:19:32.717 "ffdhe4096", 00:19:32.717 "ffdhe6144", 00:19:32.717 "ffdhe8192" 00:19:32.717 ] 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "nvmf_set_max_subsystems", 00:19:32.717 "params": { 00:19:32.717 "max_subsystems": 1024 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "nvmf_set_crdt", 00:19:32.717 "params": { 00:19:32.717 "crdt1": 0, 00:19:32.717 "crdt2": 0, 00:19:32.717 "crdt3": 0 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "nvmf_create_transport", 00:19:32.717 "params": { 00:19:32.717 "trtype": "TCP", 00:19:32.717 "max_queue_depth": 128, 00:19:32.717 "max_io_qpairs_per_ctrlr": 127, 00:19:32.717 "in_capsule_data_size": 4096, 00:19:32.717 "max_io_size": 131072, 00:19:32.717 "io_unit_size": 131072, 00:19:32.717 "max_aq_depth": 128, 00:19:32.717 "num_shared_buffers": 511, 00:19:32.717 "buf_cache_size": 4294967295, 00:19:32.717 "dif_insert_or_strip": false, 00:19:32.717 "zcopy": false, 00:19:32.717 "c2h_success": false, 00:19:32.717 "sock_priority": 0, 00:19:32.717 "abort_timeout_sec": 1, 00:19:32.717 "ack_timeout": 0, 00:19:32.717 "data_wr_pool_size": 0 00:19:32.717 } 00:19:32.717 }, 00:19:32.717 { 00:19:32.717 "method": "nvmf_create_subsystem", 00:19:32.717 "params": { 00:19:32.717 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.717 "allow_any_host": false, 00:19:32.718 "serial_number": "SPDK00000000000001", 00:19:32.718 "model_number": "SPDK bdev Controller", 00:19:32.718 "max_namespaces": 10, 00:19:32.718 "min_cntlid": 1, 00:19:32.718 "max_cntlid": 65519, 00:19:32.718 "ana_reporting": false 00:19:32.718 } 00:19:32.718 }, 00:19:32.718 { 00:19:32.718 "method": "nvmf_subsystem_add_host", 00:19:32.718 "params": { 00:19:32.718 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.718 "host": "nqn.2016-06.io.spdk:host1", 00:19:32.718 "psk": "key0" 00:19:32.718 } 00:19:32.718 }, 00:19:32.718 { 00:19:32.718 "method": "nvmf_subsystem_add_ns", 00:19:32.718 "params": { 00:19:32.718 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.718 "namespace": { 00:19:32.718 "nsid": 1, 00:19:32.718 "bdev_name": "malloc0", 00:19:32.718 "nguid": "89C813612D9846338857F7878C154989", 00:19:32.718 "uuid": "89c81361-2d98-4633-8857-f7878c154989", 00:19:32.718 "no_auto_visible": false 00:19:32.718 } 00:19:32.718 } 00:19:32.718 }, 00:19:32.718 { 00:19:32.718 "method": "nvmf_subsystem_add_listener", 00:19:32.718 "params": { 00:19:32.718 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:32.718 "listen_address": { 00:19:32.718 "trtype": "TCP", 00:19:32.718 "adrfam": "IPv4", 00:19:32.718 "traddr": "10.0.0.2", 00:19:32.718 "trsvcid": "4420" 00:19:32.718 }, 00:19:32.718 "secure_channel": true 00:19:32.718 } 00:19:32.718 } 00:19:32.718 ] 00:19:32.718 } 00:19:32.718 ] 00:19:32.718 }' 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=69312 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 69312 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 -c /dev/fd/62 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 69312 ']' 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:32.718 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:32.718 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:32.978 [2024-10-17 17:40:40.639807] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:32.978 [2024-10-17 17:40:40.639862] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:32.978 [2024-10-17 17:40:40.723614] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:32.978 [2024-10-17 17:40:40.751713] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:32.978 [2024-10-17 17:40:40.751742] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:32.978 [2024-10-17 17:40:40.751751] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:32.978 [2024-10-17 17:40:40.751755] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:32.978 [2024-10-17 17:40:40.751759] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:32.978 [2024-10-17 17:40:40.752271] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:33.237 [2024-10-17 17:40:40.945085] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:33.237 [2024-10-17 17:40:40.977106] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:33.237 [2024-10-17 17:40:40.977310] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@209 -- # bdevperf_pid=69352 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@210 -- # waitforlisten 69352 /var/tmp/bdevperf.sock 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 69352 ']' 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:33.807 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 -c /dev/fd/63 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:33.807 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # echo '{ 00:19:33.807 "subsystems": [ 00:19:33.807 { 00:19:33.807 "subsystem": "keyring", 00:19:33.807 "config": [ 00:19:33.807 { 00:19:33.807 "method": "keyring_file_add_key", 00:19:33.807 "params": { 00:19:33.807 "name": "key0", 00:19:33.807 "path": "/tmp/tmp.wGvnMMapw5" 00:19:33.807 } 00:19:33.807 } 00:19:33.807 ] 00:19:33.807 }, 00:19:33.807 { 00:19:33.807 "subsystem": "iobuf", 00:19:33.807 "config": [ 00:19:33.807 { 00:19:33.807 "method": "iobuf_set_options", 00:19:33.807 "params": { 00:19:33.808 "small_pool_count": 8192, 00:19:33.808 "large_pool_count": 1024, 00:19:33.808 "small_bufsize": 8192, 00:19:33.808 "large_bufsize": 135168 00:19:33.808 } 00:19:33.808 } 00:19:33.808 ] 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "subsystem": "sock", 00:19:33.808 "config": [ 00:19:33.808 { 00:19:33.808 "method": "sock_set_default_impl", 00:19:33.808 "params": { 00:19:33.808 "impl_name": "posix" 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "sock_impl_set_options", 00:19:33.808 "params": { 00:19:33.808 "impl_name": "ssl", 00:19:33.808 "recv_buf_size": 4096, 00:19:33.808 "send_buf_size": 4096, 00:19:33.808 "enable_recv_pipe": true, 00:19:33.808 "enable_quickack": false, 00:19:33.808 "enable_placement_id": 0, 00:19:33.808 "enable_zerocopy_send_server": true, 00:19:33.808 "enable_zerocopy_send_client": false, 00:19:33.808 "zerocopy_threshold": 0, 00:19:33.808 "tls_version": 0, 00:19:33.808 "enable_ktls": false 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "sock_impl_set_options", 00:19:33.808 "params": { 00:19:33.808 "impl_name": "posix", 00:19:33.808 "recv_buf_size": 2097152, 00:19:33.808 "send_buf_size": 2097152, 00:19:33.808 "enable_recv_pipe": true, 00:19:33.808 "enable_quickack": false, 00:19:33.808 "enable_placement_id": 0, 00:19:33.808 "enable_zerocopy_send_server": true, 00:19:33.808 "enable_zerocopy_send_client": false, 00:19:33.808 "zerocopy_threshold": 0, 00:19:33.808 "tls_version": 0, 00:19:33.808 "enable_ktls": false 00:19:33.808 } 00:19:33.808 } 00:19:33.808 ] 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "subsystem": "vmd", 00:19:33.808 "config": [] 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "subsystem": "accel", 00:19:33.808 "config": [ 00:19:33.808 { 00:19:33.808 "method": "accel_set_options", 00:19:33.808 "params": { 00:19:33.808 "small_cache_size": 128, 00:19:33.808 "large_cache_size": 16, 00:19:33.808 "task_count": 2048, 00:19:33.808 "sequence_count": 2048, 00:19:33.808 "buf_count": 2048 00:19:33.808 } 00:19:33.808 } 00:19:33.808 ] 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "subsystem": "bdev", 00:19:33.808 "config": [ 00:19:33.808 { 00:19:33.808 "method": "bdev_set_options", 00:19:33.808 "params": { 00:19:33.808 "bdev_io_pool_size": 65535, 00:19:33.808 "bdev_io_cache_size": 256, 00:19:33.808 "bdev_auto_examine": true, 00:19:33.808 "iobuf_small_cache_size": 128, 00:19:33.808 "iobuf_large_cache_size": 16 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "bdev_raid_set_options", 00:19:33.808 "params": { 00:19:33.808 "process_window_size_kb": 1024, 00:19:33.808 "process_max_bandwidth_mb_sec": 0 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "bdev_iscsi_set_options", 00:19:33.808 "params": { 00:19:33.808 "timeout_sec": 30 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "bdev_nvme_set_options", 00:19:33.808 "params": { 00:19:33.808 "action_on_timeout": "none", 00:19:33.808 "timeout_us": 0, 00:19:33.808 "timeout_admin_us": 0, 00:19:33.808 "keep_alive_timeout_ms": 10000, 00:19:33.808 "arbitration_burst": 0, 00:19:33.808 "low_priority_weight": 0, 00:19:33.808 "medium_priority_weight": 0, 00:19:33.808 "high_priority_weight": 0, 00:19:33.808 "nvme_adminq_poll_period_us": 10000, 00:19:33.808 "nvme_ioq_poll_period_us": 0, 00:19:33.808 "io_queue_requests": 512, 00:19:33.808 "delay_cmd_submit": true, 00:19:33.808 "transport_retry_count": 4, 00:19:33.808 "bdev_retry_count": 3, 00:19:33.808 "transport_ack_timeout": 0, 00:19:33.808 "ctrlr_loss_timeout_sec": 0, 00:19:33.808 "reconnect_delay_sec": 0, 00:19:33.808 "fast_io_fail_timeout_sec": 0, 00:19:33.808 "disable_auto_failback": false, 00:19:33.808 "generate_uuids": false, 00:19:33.808 "transport_tos": 0, 00:19:33.808 "nvme_error_stat": false, 00:19:33.808 "rdma_srq_size": 0, 00:19:33.808 "io_path_stat": false, 00:19:33.808 "allow_accel_sequence": false, 00:19:33.808 "rdma_max_cq_size": 0, 00:19:33.808 "rdma_cm_event_timeout_ms": 0, 00:19:33.808 "dhchap_digests": [ 00:19:33.808 "sha256", 00:19:33.808 "sha384", 00:19:33.808 "sha512" 00:19:33.808 ], 00:19:33.808 "dhchap_dhgroups": [ 00:19:33.808 "null", 00:19:33.808 "ffdhe2048", 00:19:33.808 "ffdhe3072", 00:19:33.808 "ffdhe4096", 00:19:33.808 "ffdhe6144", 00:19:33.808 "ffdhe8192" 00:19:33.808 ], 00:19:33.808 "rdma_umr_per_io": false 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "bdev_nvme_attach_controller", 00:19:33.808 "params": { 00:19:33.808 "name": "TLSTEST", 00:19:33.808 "trtype": "TCP", 00:19:33.808 "adrfam": "IPv4", 00:19:33.808 "traddr": "10.0.0.2", 00:19:33.808 "trsvcid": "4420", 00:19:33.808 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:33.808 "prchk_reftag": false, 00:19:33.808 "prchk_guard": false, 00:19:33.808 "ctrlr_loss_timeout_sec": 0, 00:19:33.808 "reconnect_delay_sec": 0, 00:19:33.808 "fast_io_fail_timeout_sec": 0, 00:19:33.808 "psk": "key0", 00:19:33.808 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:33.808 "hdgst": false, 00:19:33.808 "ddgst": false, 00:19:33.808 "multipath": "multipath" 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "bdev_nvme_set_hotplug", 00:19:33.808 "params": { 00:19:33.808 "period_us": 100000, 00:19:33.808 "enable": false 00:19:33.808 } 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "method": "bdev_wait_for_examine" 00:19:33.808 } 00:19:33.808 ] 00:19:33.808 }, 00:19:33.808 { 00:19:33.808 "subsystem": "nbd", 00:19:33.808 "config": [] 00:19:33.808 } 00:19:33.808 ] 00:19:33.808 }' 00:19:33.808 [2024-10-17 17:40:41.516708] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:33.808 [2024-10-17 17:40:41.516764] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69352 ] 00:19:33.808 [2024-10-17 17:40:41.596980] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:33.808 [2024-10-17 17:40:41.632342] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:34.069 [2024-10-17 17:40:41.771759] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:34.639 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:34.639 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:34.639 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@213 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:19:34.639 Running I/O for 10 seconds... 00:19:36.518 5985.00 IOPS, 23.38 MiB/s [2024-10-17T15:40:45.818Z] 5630.00 IOPS, 21.99 MiB/s [2024-10-17T15:40:46.756Z] 5506.33 IOPS, 21.51 MiB/s [2024-10-17T15:40:47.695Z] 5557.00 IOPS, 21.71 MiB/s [2024-10-17T15:40:48.634Z] 5660.40 IOPS, 22.11 MiB/s [2024-10-17T15:40:49.574Z] 5552.00 IOPS, 21.69 MiB/s [2024-10-17T15:40:50.514Z] 5584.57 IOPS, 21.81 MiB/s [2024-10-17T15:40:51.454Z] 5565.00 IOPS, 21.74 MiB/s [2024-10-17T15:40:52.835Z] 5604.00 IOPS, 21.89 MiB/s [2024-10-17T15:40:52.835Z] 5616.10 IOPS, 21.94 MiB/s 00:19:44.916 Latency(us) 00:19:44.916 [2024-10-17T15:40:52.835Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:44.916 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:44.916 Verification LBA range: start 0x0 length 0x2000 00:19:44.916 TLSTESTn1 : 10.01 5621.05 21.96 0.00 0.00 22739.07 5379.41 59419.31 00:19:44.916 [2024-10-17T15:40:52.835Z] =================================================================================================================== 00:19:44.916 [2024-10-17T15:40:52.835Z] Total : 5621.05 21.96 0.00 0.00 22739.07 5379.41 59419.31 00:19:44.916 { 00:19:44.916 "results": [ 00:19:44.916 { 00:19:44.916 "job": "TLSTESTn1", 00:19:44.916 "core_mask": "0x4", 00:19:44.916 "workload": "verify", 00:19:44.916 "status": "finished", 00:19:44.916 "verify_range": { 00:19:44.916 "start": 0, 00:19:44.916 "length": 8192 00:19:44.916 }, 00:19:44.916 "queue_depth": 128, 00:19:44.916 "io_size": 4096, 00:19:44.916 "runtime": 10.013793, 00:19:44.916 "iops": 5621.046890024589, 00:19:44.916 "mibps": 21.95721441415855, 00:19:44.916 "io_failed": 0, 00:19:44.916 "io_timeout": 0, 00:19:44.916 "avg_latency_us": 22739.072169793442, 00:19:44.916 "min_latency_us": 5379.413333333333, 00:19:44.916 "max_latency_us": 59419.306666666664 00:19:44.916 } 00:19:44.916 ], 00:19:44.916 "core_count": 1 00:19:44.916 } 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@215 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@216 -- # killprocess 69352 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 69352 ']' 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 69352 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69352 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69352' 00:19:44.916 killing process with pid 69352 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 69352 00:19:44.916 Received shutdown signal, test time was about 10.000000 seconds 00:19:44.916 00:19:44.916 Latency(us) 00:19:44.916 [2024-10-17T15:40:52.835Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:44.916 [2024-10-17T15:40:52.835Z] =================================================================================================================== 00:19:44.916 [2024-10-17T15:40:52.835Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 69352 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@217 -- # killprocess 69312 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 69312 ']' 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 69312 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69312 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69312' 00:19:44.916 killing process with pid 69312 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 69312 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 69312 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@220 -- # nvmfappstart 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=71683 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 71683 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 71683 ']' 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:44.916 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:44.916 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:45.177 [2024-10-17 17:40:52.848910] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:45.177 [2024-10-17 17:40:52.848964] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:45.177 [2024-10-17 17:40:52.933634] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:45.177 [2024-10-17 17:40:52.981688] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:45.177 [2024-10-17 17:40:52.981743] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:45.177 [2024-10-17 17:40:52.981752] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:45.177 [2024-10-17 17:40:52.981759] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:45.177 [2024-10-17 17:40:52.981765] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:45.177 [2024-10-17 17:40:52.982576] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:45.748 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:45.748 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:45.748 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:19:45.748 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:19:45.748 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:46.009 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:46.009 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@221 -- # setup_nvmf_tgt /tmp/tmp.wGvnMMapw5 00:19:46.009 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.wGvnMMapw5 00:19:46.009 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:46.009 [2024-10-17 17:40:53.866502] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:46.009 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:46.269 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:46.530 [2024-10-17 17:40:54.219393] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:46.530 [2024-10-17 17:40:54.219757] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:46.530 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:46.530 malloc0 00:19:46.530 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:46.790 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@224 -- # bdevperf_pid=72048 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@226 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@222 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@227 -- # waitforlisten 72048 /var/tmp/bdevperf.sock 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 72048 ']' 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:47.051 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:47.051 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:47.311 [2024-10-17 17:40:55.013072] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:47.311 [2024-10-17 17:40:55.013142] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72048 ] 00:19:47.311 [2024-10-17 17:40:55.095613] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:47.311 [2024-10-17 17:40:55.130831] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:47.881 17:40:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:47.881 17:40:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:47.881 17:40:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@229 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:48.141 17:40:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@230 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:19:48.401 [2024-10-17 17:40:56.116695] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:48.401 nvme0n1 00:19:48.401 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@234 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:48.401 Running I/O for 1 seconds... 00:19:49.782 5215.00 IOPS, 20.37 MiB/s 00:19:49.782 Latency(us) 00:19:49.782 [2024-10-17T15:40:57.701Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:49.782 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:49.782 Verification LBA range: start 0x0 length 0x2000 00:19:49.782 nvme0n1 : 1.02 5259.40 20.54 0.00 0.00 24171.82 5597.87 32768.00 00:19:49.782 [2024-10-17T15:40:57.701Z] =================================================================================================================== 00:19:49.782 [2024-10-17T15:40:57.701Z] Total : 5259.40 20.54 0.00 0.00 24171.82 5597.87 32768.00 00:19:49.782 { 00:19:49.782 "results": [ 00:19:49.782 { 00:19:49.782 "job": "nvme0n1", 00:19:49.782 "core_mask": "0x2", 00:19:49.782 "workload": "verify", 00:19:49.782 "status": "finished", 00:19:49.782 "verify_range": { 00:19:49.782 "start": 0, 00:19:49.782 "length": 8192 00:19:49.782 }, 00:19:49.782 "queue_depth": 128, 00:19:49.782 "io_size": 4096, 00:19:49.782 "runtime": 1.016086, 00:19:49.782 "iops": 5259.397334477593, 00:19:49.782 "mibps": 20.5445208378031, 00:19:49.782 "io_failed": 0, 00:19:49.782 "io_timeout": 0, 00:19:49.782 "avg_latency_us": 24171.82147704591, 00:19:49.782 "min_latency_us": 5597.866666666667, 00:19:49.782 "max_latency_us": 32768.0 00:19:49.782 } 00:19:49.782 ], 00:19:49.782 "core_count": 1 00:19:49.782 } 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@236 -- # killprocess 72048 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 72048 ']' 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 72048 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72048 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72048' 00:19:49.782 killing process with pid 72048 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 72048 00:19:49.782 Received shutdown signal, test time was about 1.000000 seconds 00:19:49.782 00:19:49.782 Latency(us) 00:19:49.782 [2024-10-17T15:40:57.701Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:49.782 [2024-10-17T15:40:57.701Z] =================================================================================================================== 00:19:49.782 [2024-10-17T15:40:57.701Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 72048 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@237 -- # killprocess 71683 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 71683 ']' 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 71683 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71683 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71683' 00:19:49.782 killing process with pid 71683 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 71683 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 71683 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@242 -- # nvmfappstart 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:19:49.782 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=72658 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 72658 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 72658 ']' 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:50.043 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:50.043 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:50.043 [2024-10-17 17:40:57.763128] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:50.043 [2024-10-17 17:40:57.763198] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:50.043 [2024-10-17 17:40:57.855743] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:50.043 [2024-10-17 17:40:57.906367] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:50.043 [2024-10-17 17:40:57.906416] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:50.043 [2024-10-17 17:40:57.906425] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:50.043 [2024-10-17 17:40:57.906432] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:50.043 [2024-10-17 17:40:57.906438] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:50.043 [2024-10-17 17:40:57.907242] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@243 -- # rpc_cmd 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:50.984 [2024-10-17 17:40:58.613263] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:50.984 malloc0 00:19:50.984 [2024-10-17 17:40:58.643407] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:50.984 [2024-10-17 17:40:58.643757] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@256 -- # bdevperf_pid=72758 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@258 -- # waitforlisten 72758 /var/tmp/bdevperf.sock 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@254 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 72758 ']' 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:50.984 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:50.984 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:50.984 [2024-10-17 17:40:58.726218] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:50.984 [2024-10-17 17:40:58.726292] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72758 ] 00:19:50.984 [2024-10-17 17:40:58.807071] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:50.984 [2024-10-17 17:40:58.842101] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:51.971 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:51.971 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:51.971 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@259 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.wGvnMMapw5 00:19:51.971 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@260 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:19:51.971 [2024-10-17 17:40:59.808618] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:51.971 nvme0n1 00:19:52.231 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@264 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:52.231 Running I/O for 1 seconds... 00:19:53.170 6051.00 IOPS, 23.64 MiB/s 00:19:53.170 Latency(us) 00:19:53.170 [2024-10-17T15:41:01.089Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:53.170 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:53.170 Verification LBA range: start 0x0 length 0x2000 00:19:53.170 nvme0n1 : 1.01 6107.42 23.86 0.00 0.00 20830.45 4997.12 35170.99 00:19:53.170 [2024-10-17T15:41:01.089Z] =================================================================================================================== 00:19:53.170 [2024-10-17T15:41:01.089Z] Total : 6107.42 23.86 0.00 0.00 20830.45 4997.12 35170.99 00:19:53.170 { 00:19:53.170 "results": [ 00:19:53.170 { 00:19:53.170 "job": "nvme0n1", 00:19:53.170 "core_mask": "0x2", 00:19:53.170 "workload": "verify", 00:19:53.170 "status": "finished", 00:19:53.170 "verify_range": { 00:19:53.170 "start": 0, 00:19:53.170 "length": 8192 00:19:53.170 }, 00:19:53.170 "queue_depth": 128, 00:19:53.170 "io_size": 4096, 00:19:53.170 "runtime": 1.01172, 00:19:53.170 "iops": 6107.4210255802, 00:19:53.170 "mibps": 23.857113381172656, 00:19:53.170 "io_failed": 0, 00:19:53.170 "io_timeout": 0, 00:19:53.170 "avg_latency_us": 20830.452496088907, 00:19:53.170 "min_latency_us": 4997.12, 00:19:53.171 "max_latency_us": 35170.986666666664 00:19:53.171 } 00:19:53.171 ], 00:19:53.171 "core_count": 1 00:19:53.171 } 00:19:53.171 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # rpc_cmd save_config 00:19:53.171 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:53.171 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:53.431 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:53.431 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # tgtcfg='{ 00:19:53.431 "subsystems": [ 00:19:53.431 { 00:19:53.431 "subsystem": "keyring", 00:19:53.431 "config": [ 00:19:53.431 { 00:19:53.431 "method": "keyring_file_add_key", 00:19:53.431 "params": { 00:19:53.431 "name": "key0", 00:19:53.431 "path": "/tmp/tmp.wGvnMMapw5" 00:19:53.431 } 00:19:53.431 } 00:19:53.431 ] 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "subsystem": "iobuf", 00:19:53.431 "config": [ 00:19:53.431 { 00:19:53.431 "method": "iobuf_set_options", 00:19:53.431 "params": { 00:19:53.431 "small_pool_count": 8192, 00:19:53.431 "large_pool_count": 1024, 00:19:53.431 "small_bufsize": 8192, 00:19:53.431 "large_bufsize": 135168 00:19:53.431 } 00:19:53.431 } 00:19:53.431 ] 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "subsystem": "sock", 00:19:53.431 "config": [ 00:19:53.431 { 00:19:53.431 "method": "sock_set_default_impl", 00:19:53.431 "params": { 00:19:53.431 "impl_name": "posix" 00:19:53.431 } 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "method": "sock_impl_set_options", 00:19:53.431 "params": { 00:19:53.431 "impl_name": "ssl", 00:19:53.431 "recv_buf_size": 4096, 00:19:53.431 "send_buf_size": 4096, 00:19:53.431 "enable_recv_pipe": true, 00:19:53.431 "enable_quickack": false, 00:19:53.431 "enable_placement_id": 0, 00:19:53.431 "enable_zerocopy_send_server": true, 00:19:53.431 "enable_zerocopy_send_client": false, 00:19:53.431 "zerocopy_threshold": 0, 00:19:53.431 "tls_version": 0, 00:19:53.431 "enable_ktls": false 00:19:53.431 } 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "method": "sock_impl_set_options", 00:19:53.431 "params": { 00:19:53.431 "impl_name": "posix", 00:19:53.431 "recv_buf_size": 2097152, 00:19:53.431 "send_buf_size": 2097152, 00:19:53.431 "enable_recv_pipe": true, 00:19:53.431 "enable_quickack": false, 00:19:53.431 "enable_placement_id": 0, 00:19:53.431 "enable_zerocopy_send_server": true, 00:19:53.431 "enable_zerocopy_send_client": false, 00:19:53.431 "zerocopy_threshold": 0, 00:19:53.431 "tls_version": 0, 00:19:53.431 "enable_ktls": false 00:19:53.431 } 00:19:53.431 } 00:19:53.431 ] 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "subsystem": "vmd", 00:19:53.431 "config": [] 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "subsystem": "accel", 00:19:53.431 "config": [ 00:19:53.431 { 00:19:53.431 "method": "accel_set_options", 00:19:53.431 "params": { 00:19:53.431 "small_cache_size": 128, 00:19:53.431 "large_cache_size": 16, 00:19:53.431 "task_count": 2048, 00:19:53.431 "sequence_count": 2048, 00:19:53.431 "buf_count": 2048 00:19:53.431 } 00:19:53.431 } 00:19:53.431 ] 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "subsystem": "bdev", 00:19:53.431 "config": [ 00:19:53.431 { 00:19:53.431 "method": "bdev_set_options", 00:19:53.431 "params": { 00:19:53.431 "bdev_io_pool_size": 65535, 00:19:53.431 "bdev_io_cache_size": 256, 00:19:53.431 "bdev_auto_examine": true, 00:19:53.431 "iobuf_small_cache_size": 128, 00:19:53.431 "iobuf_large_cache_size": 16 00:19:53.431 } 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "method": "bdev_raid_set_options", 00:19:53.431 "params": { 00:19:53.431 "process_window_size_kb": 1024, 00:19:53.431 "process_max_bandwidth_mb_sec": 0 00:19:53.431 } 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "method": "bdev_iscsi_set_options", 00:19:53.431 "params": { 00:19:53.431 "timeout_sec": 30 00:19:53.431 } 00:19:53.431 }, 00:19:53.431 { 00:19:53.431 "method": "bdev_nvme_set_options", 00:19:53.431 "params": { 00:19:53.431 "action_on_timeout": "none", 00:19:53.431 "timeout_us": 0, 00:19:53.431 "timeout_admin_us": 0, 00:19:53.431 "keep_alive_timeout_ms": 10000, 00:19:53.431 "arbitration_burst": 0, 00:19:53.431 "low_priority_weight": 0, 00:19:53.432 "medium_priority_weight": 0, 00:19:53.432 "high_priority_weight": 0, 00:19:53.432 "nvme_adminq_poll_period_us": 10000, 00:19:53.432 "nvme_ioq_poll_period_us": 0, 00:19:53.432 "io_queue_requests": 0, 00:19:53.432 "delay_cmd_submit": true, 00:19:53.432 "transport_retry_count": 4, 00:19:53.432 "bdev_retry_count": 3, 00:19:53.432 "transport_ack_timeout": 0, 00:19:53.432 "ctrlr_loss_timeout_sec": 0, 00:19:53.432 "reconnect_delay_sec": 0, 00:19:53.432 "fast_io_fail_timeout_sec": 0, 00:19:53.432 "disable_auto_failback": false, 00:19:53.432 "generate_uuids": false, 00:19:53.432 "transport_tos": 0, 00:19:53.432 "nvme_error_stat": false, 00:19:53.432 "rdma_srq_size": 0, 00:19:53.432 "io_path_stat": false, 00:19:53.432 "allow_accel_sequence": false, 00:19:53.432 "rdma_max_cq_size": 0, 00:19:53.432 "rdma_cm_event_timeout_ms": 0, 00:19:53.432 "dhchap_digests": [ 00:19:53.432 "sha256", 00:19:53.432 "sha384", 00:19:53.432 "sha512" 00:19:53.432 ], 00:19:53.432 "dhchap_dhgroups": [ 00:19:53.432 "null", 00:19:53.432 "ffdhe2048", 00:19:53.432 "ffdhe3072", 00:19:53.432 "ffdhe4096", 00:19:53.432 "ffdhe6144", 00:19:53.432 "ffdhe8192" 00:19:53.432 ], 00:19:53.432 "rdma_umr_per_io": false 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "bdev_nvme_set_hotplug", 00:19:53.432 "params": { 00:19:53.432 "period_us": 100000, 00:19:53.432 "enable": false 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "bdev_malloc_create", 00:19:53.432 "params": { 00:19:53.432 "name": "malloc0", 00:19:53.432 "num_blocks": 8192, 00:19:53.432 "block_size": 4096, 00:19:53.432 "physical_block_size": 4096, 00:19:53.432 "uuid": "5177791b-37f6-4e9f-bcce-1d070c9a5645", 00:19:53.432 "optimal_io_boundary": 0, 00:19:53.432 "md_size": 0, 00:19:53.432 "dif_type": 0, 00:19:53.432 "dif_is_head_of_md": false, 00:19:53.432 "dif_pi_format": 0 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "bdev_wait_for_examine" 00:19:53.432 } 00:19:53.432 ] 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "subsystem": "nbd", 00:19:53.432 "config": [] 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "subsystem": "scheduler", 00:19:53.432 "config": [ 00:19:53.432 { 00:19:53.432 "method": "framework_set_scheduler", 00:19:53.432 "params": { 00:19:53.432 "name": "static" 00:19:53.432 } 00:19:53.432 } 00:19:53.432 ] 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "subsystem": "nvmf", 00:19:53.432 "config": [ 00:19:53.432 { 00:19:53.432 "method": "nvmf_set_config", 00:19:53.432 "params": { 00:19:53.432 "discovery_filter": "match_any", 00:19:53.432 "admin_cmd_passthru": { 00:19:53.432 "identify_ctrlr": false 00:19:53.432 }, 00:19:53.432 "dhchap_digests": [ 00:19:53.432 "sha256", 00:19:53.432 "sha384", 00:19:53.432 "sha512" 00:19:53.432 ], 00:19:53.432 "dhchap_dhgroups": [ 00:19:53.432 "null", 00:19:53.432 "ffdhe2048", 00:19:53.432 "ffdhe3072", 00:19:53.432 "ffdhe4096", 00:19:53.432 "ffdhe6144", 00:19:53.432 "ffdhe8192" 00:19:53.432 ] 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "nvmf_set_max_subsystems", 00:19:53.432 "params": { 00:19:53.432 "max_subsystems": 1024 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "nvmf_set_crdt", 00:19:53.432 "params": { 00:19:53.432 "crdt1": 0, 00:19:53.432 "crdt2": 0, 00:19:53.432 "crdt3": 0 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "nvmf_create_transport", 00:19:53.432 "params": { 00:19:53.432 "trtype": "TCP", 00:19:53.432 "max_queue_depth": 128, 00:19:53.432 "max_io_qpairs_per_ctrlr": 127, 00:19:53.432 "in_capsule_data_size": 4096, 00:19:53.432 "max_io_size": 131072, 00:19:53.432 "io_unit_size": 131072, 00:19:53.432 "max_aq_depth": 128, 00:19:53.432 "num_shared_buffers": 511, 00:19:53.432 "buf_cache_size": 4294967295, 00:19:53.432 "dif_insert_or_strip": false, 00:19:53.432 "zcopy": false, 00:19:53.432 "c2h_success": false, 00:19:53.432 "sock_priority": 0, 00:19:53.432 "abort_timeout_sec": 1, 00:19:53.432 "ack_timeout": 0, 00:19:53.432 "data_wr_pool_size": 0 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "nvmf_create_subsystem", 00:19:53.432 "params": { 00:19:53.432 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.432 "allow_any_host": false, 00:19:53.432 "serial_number": "00000000000000000000", 00:19:53.432 "model_number": "SPDK bdev Controller", 00:19:53.432 "max_namespaces": 32, 00:19:53.432 "min_cntlid": 1, 00:19:53.432 "max_cntlid": 65519, 00:19:53.432 "ana_reporting": false 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "nvmf_subsystem_add_host", 00:19:53.432 "params": { 00:19:53.432 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.432 "host": "nqn.2016-06.io.spdk:host1", 00:19:53.432 "psk": "key0" 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "nvmf_subsystem_add_ns", 00:19:53.432 "params": { 00:19:53.432 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.432 "namespace": { 00:19:53.432 "nsid": 1, 00:19:53.432 "bdev_name": "malloc0", 00:19:53.432 "nguid": "5177791B37F64E9FBCCE1D070C9A5645", 00:19:53.432 "uuid": "5177791b-37f6-4e9f-bcce-1d070c9a5645", 00:19:53.432 "no_auto_visible": false 00:19:53.432 } 00:19:53.432 } 00:19:53.432 }, 00:19:53.432 { 00:19:53.432 "method": "nvmf_subsystem_add_listener", 00:19:53.432 "params": { 00:19:53.432 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.432 "listen_address": { 00:19:53.432 "trtype": "TCP", 00:19:53.432 "adrfam": "IPv4", 00:19:53.432 "traddr": "10.0.0.2", 00:19:53.432 "trsvcid": "4420" 00:19:53.432 }, 00:19:53.432 "secure_channel": false, 00:19:53.432 "sock_impl": "ssl" 00:19:53.432 } 00:19:53.432 } 00:19:53.432 ] 00:19:53.432 } 00:19:53.432 ] 00:19:53.432 }' 00:19:53.432 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # bperfcfg='{ 00:19:53.693 "subsystems": [ 00:19:53.693 { 00:19:53.693 "subsystem": "keyring", 00:19:53.693 "config": [ 00:19:53.693 { 00:19:53.693 "method": "keyring_file_add_key", 00:19:53.693 "params": { 00:19:53.693 "name": "key0", 00:19:53.693 "path": "/tmp/tmp.wGvnMMapw5" 00:19:53.693 } 00:19:53.693 } 00:19:53.693 ] 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "subsystem": "iobuf", 00:19:53.693 "config": [ 00:19:53.693 { 00:19:53.693 "method": "iobuf_set_options", 00:19:53.693 "params": { 00:19:53.693 "small_pool_count": 8192, 00:19:53.693 "large_pool_count": 1024, 00:19:53.693 "small_bufsize": 8192, 00:19:53.693 "large_bufsize": 135168 00:19:53.693 } 00:19:53.693 } 00:19:53.693 ] 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "subsystem": "sock", 00:19:53.693 "config": [ 00:19:53.693 { 00:19:53.693 "method": "sock_set_default_impl", 00:19:53.693 "params": { 00:19:53.693 "impl_name": "posix" 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "sock_impl_set_options", 00:19:53.693 "params": { 00:19:53.693 "impl_name": "ssl", 00:19:53.693 "recv_buf_size": 4096, 00:19:53.693 "send_buf_size": 4096, 00:19:53.693 "enable_recv_pipe": true, 00:19:53.693 "enable_quickack": false, 00:19:53.693 "enable_placement_id": 0, 00:19:53.693 "enable_zerocopy_send_server": true, 00:19:53.693 "enable_zerocopy_send_client": false, 00:19:53.693 "zerocopy_threshold": 0, 00:19:53.693 "tls_version": 0, 00:19:53.693 "enable_ktls": false 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "sock_impl_set_options", 00:19:53.693 "params": { 00:19:53.693 "impl_name": "posix", 00:19:53.693 "recv_buf_size": 2097152, 00:19:53.693 "send_buf_size": 2097152, 00:19:53.693 "enable_recv_pipe": true, 00:19:53.693 "enable_quickack": false, 00:19:53.693 "enable_placement_id": 0, 00:19:53.693 "enable_zerocopy_send_server": true, 00:19:53.693 "enable_zerocopy_send_client": false, 00:19:53.693 "zerocopy_threshold": 0, 00:19:53.693 "tls_version": 0, 00:19:53.693 "enable_ktls": false 00:19:53.693 } 00:19:53.693 } 00:19:53.693 ] 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "subsystem": "vmd", 00:19:53.693 "config": [] 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "subsystem": "accel", 00:19:53.693 "config": [ 00:19:53.693 { 00:19:53.693 "method": "accel_set_options", 00:19:53.693 "params": { 00:19:53.693 "small_cache_size": 128, 00:19:53.693 "large_cache_size": 16, 00:19:53.693 "task_count": 2048, 00:19:53.693 "sequence_count": 2048, 00:19:53.693 "buf_count": 2048 00:19:53.693 } 00:19:53.693 } 00:19:53.693 ] 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "subsystem": "bdev", 00:19:53.693 "config": [ 00:19:53.693 { 00:19:53.693 "method": "bdev_set_options", 00:19:53.693 "params": { 00:19:53.693 "bdev_io_pool_size": 65535, 00:19:53.693 "bdev_io_cache_size": 256, 00:19:53.693 "bdev_auto_examine": true, 00:19:53.693 "iobuf_small_cache_size": 128, 00:19:53.693 "iobuf_large_cache_size": 16 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "bdev_raid_set_options", 00:19:53.693 "params": { 00:19:53.693 "process_window_size_kb": 1024, 00:19:53.693 "process_max_bandwidth_mb_sec": 0 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "bdev_iscsi_set_options", 00:19:53.693 "params": { 00:19:53.693 "timeout_sec": 30 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "bdev_nvme_set_options", 00:19:53.693 "params": { 00:19:53.693 "action_on_timeout": "none", 00:19:53.693 "timeout_us": 0, 00:19:53.693 "timeout_admin_us": 0, 00:19:53.693 "keep_alive_timeout_ms": 10000, 00:19:53.693 "arbitration_burst": 0, 00:19:53.693 "low_priority_weight": 0, 00:19:53.693 "medium_priority_weight": 0, 00:19:53.693 "high_priority_weight": 0, 00:19:53.693 "nvme_adminq_poll_period_us": 10000, 00:19:53.693 "nvme_ioq_poll_period_us": 0, 00:19:53.693 "io_queue_requests": 512, 00:19:53.693 "delay_cmd_submit": true, 00:19:53.693 "transport_retry_count": 4, 00:19:53.693 "bdev_retry_count": 3, 00:19:53.693 "transport_ack_timeout": 0, 00:19:53.693 "ctrlr_loss_timeout_sec": 0, 00:19:53.693 "reconnect_delay_sec": 0, 00:19:53.693 "fast_io_fail_timeout_sec": 0, 00:19:53.693 "disable_auto_failback": false, 00:19:53.693 "generate_uuids": false, 00:19:53.693 "transport_tos": 0, 00:19:53.693 "nvme_error_stat": false, 00:19:53.693 "rdma_srq_size": 0, 00:19:53.693 "io_path_stat": false, 00:19:53.693 "allow_accel_sequence": false, 00:19:53.693 "rdma_max_cq_size": 0, 00:19:53.693 "rdma_cm_event_timeout_ms": 0, 00:19:53.693 "dhchap_digests": [ 00:19:53.693 "sha256", 00:19:53.693 "sha384", 00:19:53.693 "sha512" 00:19:53.693 ], 00:19:53.693 "dhchap_dhgroups": [ 00:19:53.693 "null", 00:19:53.693 "ffdhe2048", 00:19:53.693 "ffdhe3072", 00:19:53.693 "ffdhe4096", 00:19:53.693 "ffdhe6144", 00:19:53.693 "ffdhe8192" 00:19:53.693 ], 00:19:53.693 "rdma_umr_per_io": false 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "bdev_nvme_attach_controller", 00:19:53.693 "params": { 00:19:53.693 "name": "nvme0", 00:19:53.693 "trtype": "TCP", 00:19:53.693 "adrfam": "IPv4", 00:19:53.693 "traddr": "10.0.0.2", 00:19:53.693 "trsvcid": "4420", 00:19:53.693 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.693 "prchk_reftag": false, 00:19:53.693 "prchk_guard": false, 00:19:53.693 "ctrlr_loss_timeout_sec": 0, 00:19:53.693 "reconnect_delay_sec": 0, 00:19:53.693 "fast_io_fail_timeout_sec": 0, 00:19:53.693 "psk": "key0", 00:19:53.693 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:53.693 "hdgst": false, 00:19:53.693 "ddgst": false, 00:19:53.693 "multipath": "multipath" 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "bdev_nvme_set_hotplug", 00:19:53.693 "params": { 00:19:53.693 "period_us": 100000, 00:19:53.693 "enable": false 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "bdev_enable_histogram", 00:19:53.693 "params": { 00:19:53.693 "name": "nvme0n1", 00:19:53.693 "enable": true 00:19:53.693 } 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "method": "bdev_wait_for_examine" 00:19:53.693 } 00:19:53.693 ] 00:19:53.693 }, 00:19:53.693 { 00:19:53.693 "subsystem": "nbd", 00:19:53.693 "config": [] 00:19:53.693 } 00:19:53.693 ] 00:19:53.693 }' 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@270 -- # killprocess 72758 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 72758 ']' 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 72758 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72758 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72758' 00:19:53.693 killing process with pid 72758 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 72758 00:19:53.693 Received shutdown signal, test time was about 1.000000 seconds 00:19:53.693 00:19:53.693 Latency(us) 00:19:53.693 [2024-10-17T15:41:01.612Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:53.693 [2024-10-17T15:41:01.612Z] =================================================================================================================== 00:19:53.693 [2024-10-17T15:41:01.612Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 72758 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@271 -- # killprocess 72658 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 72658 ']' 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 72658 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:53.693 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72658 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72658' 00:19:53.953 killing process with pid 72658 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 72658 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 72658 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # nvmfappstart -c /dev/fd/62 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:19:53.953 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # echo '{ 00:19:53.953 "subsystems": [ 00:19:53.953 { 00:19:53.953 "subsystem": "keyring", 00:19:53.953 "config": [ 00:19:53.953 { 00:19:53.953 "method": "keyring_file_add_key", 00:19:53.953 "params": { 00:19:53.953 "name": "key0", 00:19:53.953 "path": "/tmp/tmp.wGvnMMapw5" 00:19:53.953 } 00:19:53.953 } 00:19:53.953 ] 00:19:53.953 }, 00:19:53.953 { 00:19:53.953 "subsystem": "iobuf", 00:19:53.953 "config": [ 00:19:53.953 { 00:19:53.953 "method": "iobuf_set_options", 00:19:53.953 "params": { 00:19:53.953 "small_pool_count": 8192, 00:19:53.953 "large_pool_count": 1024, 00:19:53.953 "small_bufsize": 8192, 00:19:53.953 "large_bufsize": 135168 00:19:53.953 } 00:19:53.953 } 00:19:53.953 ] 00:19:53.953 }, 00:19:53.953 { 00:19:53.953 "subsystem": "sock", 00:19:53.953 "config": [ 00:19:53.953 { 00:19:53.953 "method": "sock_set_default_impl", 00:19:53.954 "params": { 00:19:53.954 "impl_name": "posix" 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "sock_impl_set_options", 00:19:53.954 "params": { 00:19:53.954 "impl_name": "ssl", 00:19:53.954 "recv_buf_size": 4096, 00:19:53.954 "send_buf_size": 4096, 00:19:53.954 "enable_recv_pipe": true, 00:19:53.954 "enable_quickack": false, 00:19:53.954 "enable_placement_id": 0, 00:19:53.954 "enable_zerocopy_send_server": true, 00:19:53.954 "enable_zerocopy_send_client": false, 00:19:53.954 "zerocopy_threshold": 0, 00:19:53.954 "tls_version": 0, 00:19:53.954 "enable_ktls": false 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "sock_impl_set_options", 00:19:53.954 "params": { 00:19:53.954 "impl_name": "posix", 00:19:53.954 "recv_buf_size": 2097152, 00:19:53.954 "send_buf_size": 2097152, 00:19:53.954 "enable_recv_pipe": true, 00:19:53.954 "enable_quickack": false, 00:19:53.954 "enable_placement_id": 0, 00:19:53.954 "enable_zerocopy_send_server": true, 00:19:53.954 "enable_zerocopy_send_client": false, 00:19:53.954 "zerocopy_threshold": 0, 00:19:53.954 "tls_version": 0, 00:19:53.954 "enable_ktls": false 00:19:53.954 } 00:19:53.954 } 00:19:53.954 ] 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "subsystem": "vmd", 00:19:53.954 "config": [] 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "subsystem": "accel", 00:19:53.954 "config": [ 00:19:53.954 { 00:19:53.954 "method": "accel_set_options", 00:19:53.954 "params": { 00:19:53.954 "small_cache_size": 128, 00:19:53.954 "large_cache_size": 16, 00:19:53.954 "task_count": 2048, 00:19:53.954 "sequence_count": 2048, 00:19:53.954 "buf_count": 2048 00:19:53.954 } 00:19:53.954 } 00:19:53.954 ] 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "subsystem": "bdev", 00:19:53.954 "config": [ 00:19:53.954 { 00:19:53.954 "method": "bdev_set_options", 00:19:53.954 "params": { 00:19:53.954 "bdev_io_pool_size": 65535, 00:19:53.954 "bdev_io_cache_size": 256, 00:19:53.954 "bdev_auto_examine": true, 00:19:53.954 "iobuf_small_cache_size": 128, 00:19:53.954 "iobuf_large_cache_size": 16 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "bdev_raid_set_options", 00:19:53.954 "params": { 00:19:53.954 "process_window_size_kb": 1024, 00:19:53.954 "process_max_bandwidth_mb_sec": 0 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "bdev_iscsi_set_options", 00:19:53.954 "params": { 00:19:53.954 "timeout_sec": 30 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "bdev_nvme_set_options", 00:19:53.954 "params": { 00:19:53.954 "action_on_timeout": "none", 00:19:53.954 "timeout_us": 0, 00:19:53.954 "timeout_admin_us": 0, 00:19:53.954 "keep_alive_timeout_ms": 10000, 00:19:53.954 "arbitration_burst": 0, 00:19:53.954 "low_priority_weight": 0, 00:19:53.954 "medium_priority_weight": 0, 00:19:53.954 "high_priority_weight": 0, 00:19:53.954 "nvme_adminq_poll_period_us": 10000, 00:19:53.954 "nvme_ioq_poll_period_us": 0, 00:19:53.954 "io_queue_requests": 0, 00:19:53.954 "delay_cmd_submit": true, 00:19:53.954 "transport_retry_count": 4, 00:19:53.954 "bdev_retry_count": 3, 00:19:53.954 "transport_ack_timeout": 0, 00:19:53.954 "ctrlr_loss_timeout_sec": 0, 00:19:53.954 "reconnect_delay_sec": 0, 00:19:53.954 "fast_io_fail_timeout_sec": 0, 00:19:53.954 "disable_auto_failback": false, 00:19:53.954 "generate_uuids": false, 00:19:53.954 "transport_tos": 0, 00:19:53.954 "nvme_error_stat": false, 00:19:53.954 "rdma_srq_size": 0, 00:19:53.954 "io_path_stat": false, 00:19:53.954 "allow_accel_sequence": false, 00:19:53.954 "rdma_max_cq_size": 0, 00:19:53.954 "rdma_cm_event_timeout_ms": 0, 00:19:53.954 "dhchap_digests": [ 00:19:53.954 "sha256", 00:19:53.954 "sha384", 00:19:53.954 "sha512" 00:19:53.954 ], 00:19:53.954 "dhchap_dhgroups": [ 00:19:53.954 "null", 00:19:53.954 "ffdhe2048", 00:19:53.954 "ffdhe3072", 00:19:53.954 "ffdhe4096", 00:19:53.954 "ffdhe6144", 00:19:53.954 "ffdhe8192" 00:19:53.954 ], 00:19:53.954 "rdma_umr_per_io": false 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "bdev_nvme_set_hotplug", 00:19:53.954 "params": { 00:19:53.954 "period_us": 100000, 00:19:53.954 "enable": false 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "bdev_malloc_create", 00:19:53.954 "params": { 00:19:53.954 "name": "malloc0", 00:19:53.954 "num_blocks": 8192, 00:19:53.954 "block_size": 4096, 00:19:53.954 "physical_block_size": 4096, 00:19:53.954 "uuid": "5177791b-37f6-4e9f-bcce-1d070c9a5645", 00:19:53.954 "optimal_io_boundary": 0, 00:19:53.954 "md_size": 0, 00:19:53.954 "dif_type": 0, 00:19:53.954 "dif_is_head_of_md": false, 00:19:53.954 "dif_pi_format": 0 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "bdev_wait_for_examine" 00:19:53.954 } 00:19:53.954 ] 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "subsystem": "nbd", 00:19:53.954 "config": [] 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "subsystem": "scheduler", 00:19:53.954 "config": [ 00:19:53.954 { 00:19:53.954 "method": "framework_set_scheduler", 00:19:53.954 "params": { 00:19:53.954 "name": "static" 00:19:53.954 } 00:19:53.954 } 00:19:53.954 ] 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "subsystem": "nvmf", 00:19:53.954 "config": [ 00:19:53.954 { 00:19:53.954 "method": "nvmf_set_config", 00:19:53.954 "params": { 00:19:53.954 "discovery_filter": "match_any", 00:19:53.954 "admin_cmd_passthru": { 00:19:53.954 "identify_ctrlr": false 00:19:53.954 }, 00:19:53.954 "dhchap_digests": [ 00:19:53.954 "sha256", 00:19:53.954 "sha384", 00:19:53.954 "sha512" 00:19:53.954 ], 00:19:53.954 "dhchap_dhgroups": [ 00:19:53.954 "null", 00:19:53.954 "ffdhe2048", 00:19:53.954 "ffdhe3072", 00:19:53.954 "ffdhe4096", 00:19:53.954 "ffdhe6144", 00:19:53.954 "ffdhe8192" 00:19:53.954 ] 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "nvmf_set_max_subsystems", 00:19:53.954 "params": { 00:19:53.954 "max_subsystems": 1024 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "nvmf_set_crdt", 00:19:53.954 "params": { 00:19:53.954 "crdt1": 0, 00:19:53.954 "crdt2": 0, 00:19:53.954 "crdt3": 0 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "nvmf_create_transport", 00:19:53.954 "params": { 00:19:53.954 "trtype": "TCP", 00:19:53.954 "max_queue_depth": 128, 00:19:53.954 "max_io_qpairs_per_ctrlr": 127, 00:19:53.954 "in_capsule_data_size": 4096, 00:19:53.954 "max_io_size": 131072, 00:19:53.954 "io_unit_size": 131072, 00:19:53.954 "max_aq_depth": 128, 00:19:53.954 "num_shared_buffers": 511, 00:19:53.954 "buf_cache_size": 4294967295, 00:19:53.954 "dif_insert_or_strip": false, 00:19:53.954 "zcopy": false, 00:19:53.954 "c2h_success": false, 00:19:53.954 "sock_priority": 0, 00:19:53.954 "abort_timeout_sec": 1, 00:19:53.954 "ack_timeout": 0, 00:19:53.954 "data_wr_pool_size": 0 00:19:53.954 } 00:19:53.954 }, 00:19:53.954 { 00:19:53.954 "method": "nvmf_create_subsystem", 00:19:53.954 "params": { 00:19:53.954 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.955 "allow_any_host": false, 00:19:53.955 "serial_number": "00000000000000000000", 00:19:53.955 "model_number": "SPDK bdev Controller", 00:19:53.955 "max_namespaces": 32, 00:19:53.955 "min_cntlid": 1, 00:19:53.955 "max_cntlid": 65519, 00:19:53.955 "ana_reporting": false 00:19:53.955 } 00:19:53.955 }, 00:19:53.955 { 00:19:53.955 "method": "nvmf_subsystem_add_host", 00:19:53.955 "params": { 00:19:53.955 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.955 "host": "nqn.2016-06.io.spdk:host1", 00:19:53.955 "psk": "key0" 00:19:53.955 } 00:19:53.955 }, 00:19:53.955 { 00:19:53.955 "method": "nvmf_subsystem_add_ns", 00:19:53.955 "params": { 00:19:53.955 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.955 "namespace": { 00:19:53.955 "nsid": 1, 00:19:53.955 "bdev_name": "malloc0", 00:19:53.955 "nguid": "5177791B37F64E9FBCCE1D070C9A5645", 00:19:53.955 "uuid": "5177791b-37f6-4e9f-bcce-1d070c9a5645", 00:19:53.955 "no_auto_visible": false 00:19:53.955 } 00:19:53.955 } 00:19:53.955 }, 00:19:53.955 { 00:19:53.955 "method": "nvmf_subsystem_add_listener", 00:19:53.955 "params": { 00:19:53.955 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:53.955 "listen_address": { 00:19:53.955 "trtype": "TCP", 00:19:53.955 "adrfam": "IPv4", 00:19:53.955 "traddr": "10.0.0.2", 00:19:53.955 "trsvcid": "4420" 00:19:53.955 }, 00:19:53.955 "secure_channel": false, 00:19:53.955 "sock_impl": "ssl" 00:19:53.955 } 00:19:53.955 } 00:19:53.955 ] 00:19:53.955 } 00:19:53.955 ] 00:19:53.955 }' 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=73445 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 73445 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -c /dev/fd/62 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 73445 ']' 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:53.955 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:53.955 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:53.955 [2024-10-17 17:41:01.796420] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:53.955 [2024-10-17 17:41:01.796472] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:54.216 [2024-10-17 17:41:01.878749] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:54.216 [2024-10-17 17:41:01.908650] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:54.216 [2024-10-17 17:41:01.908679] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:54.216 [2024-10-17 17:41:01.908685] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:54.216 [2024-10-17 17:41:01.908695] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:54.216 [2024-10-17 17:41:01.908699] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:54.216 [2024-10-17 17:41:01.909200] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:54.216 [2024-10-17 17:41:02.102365] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:54.216 [2024-10-17 17:41:02.134393] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:54.216 [2024-10-17 17:41:02.134612] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@276 -- # bdevperf_pid=73589 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@277 -- # waitforlisten 73589 /var/tmp/bdevperf.sock 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 73589 ']' 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:54.787 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 -c /dev/fd/63 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:54.787 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # echo '{ 00:19:54.787 "subsystems": [ 00:19:54.787 { 00:19:54.787 "subsystem": "keyring", 00:19:54.787 "config": [ 00:19:54.787 { 00:19:54.787 "method": "keyring_file_add_key", 00:19:54.787 "params": { 00:19:54.787 "name": "key0", 00:19:54.787 "path": "/tmp/tmp.wGvnMMapw5" 00:19:54.787 } 00:19:54.787 } 00:19:54.787 ] 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "subsystem": "iobuf", 00:19:54.787 "config": [ 00:19:54.787 { 00:19:54.787 "method": "iobuf_set_options", 00:19:54.787 "params": { 00:19:54.787 "small_pool_count": 8192, 00:19:54.787 "large_pool_count": 1024, 00:19:54.787 "small_bufsize": 8192, 00:19:54.787 "large_bufsize": 135168 00:19:54.787 } 00:19:54.787 } 00:19:54.787 ] 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "subsystem": "sock", 00:19:54.787 "config": [ 00:19:54.787 { 00:19:54.787 "method": "sock_set_default_impl", 00:19:54.787 "params": { 00:19:54.787 "impl_name": "posix" 00:19:54.787 } 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "method": "sock_impl_set_options", 00:19:54.787 "params": { 00:19:54.787 "impl_name": "ssl", 00:19:54.787 "recv_buf_size": 4096, 00:19:54.787 "send_buf_size": 4096, 00:19:54.787 "enable_recv_pipe": true, 00:19:54.787 "enable_quickack": false, 00:19:54.787 "enable_placement_id": 0, 00:19:54.787 "enable_zerocopy_send_server": true, 00:19:54.787 "enable_zerocopy_send_client": false, 00:19:54.787 "zerocopy_threshold": 0, 00:19:54.787 "tls_version": 0, 00:19:54.787 "enable_ktls": false 00:19:54.787 } 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "method": "sock_impl_set_options", 00:19:54.787 "params": { 00:19:54.787 "impl_name": "posix", 00:19:54.787 "recv_buf_size": 2097152, 00:19:54.787 "send_buf_size": 2097152, 00:19:54.787 "enable_recv_pipe": true, 00:19:54.787 "enable_quickack": false, 00:19:54.787 "enable_placement_id": 0, 00:19:54.787 "enable_zerocopy_send_server": true, 00:19:54.787 "enable_zerocopy_send_client": false, 00:19:54.787 "zerocopy_threshold": 0, 00:19:54.787 "tls_version": 0, 00:19:54.787 "enable_ktls": false 00:19:54.787 } 00:19:54.787 } 00:19:54.787 ] 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "subsystem": "vmd", 00:19:54.787 "config": [] 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "subsystem": "accel", 00:19:54.787 "config": [ 00:19:54.787 { 00:19:54.787 "method": "accel_set_options", 00:19:54.787 "params": { 00:19:54.787 "small_cache_size": 128, 00:19:54.787 "large_cache_size": 16, 00:19:54.787 "task_count": 2048, 00:19:54.787 "sequence_count": 2048, 00:19:54.787 "buf_count": 2048 00:19:54.787 } 00:19:54.787 } 00:19:54.787 ] 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "subsystem": "bdev", 00:19:54.787 "config": [ 00:19:54.787 { 00:19:54.787 "method": "bdev_set_options", 00:19:54.787 "params": { 00:19:54.787 "bdev_io_pool_size": 65535, 00:19:54.787 "bdev_io_cache_size": 256, 00:19:54.787 "bdev_auto_examine": true, 00:19:54.787 "iobuf_small_cache_size": 128, 00:19:54.787 "iobuf_large_cache_size": 16 00:19:54.787 } 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "method": "bdev_raid_set_options", 00:19:54.787 "params": { 00:19:54.787 "process_window_size_kb": 1024, 00:19:54.787 "process_max_bandwidth_mb_sec": 0 00:19:54.787 } 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "method": "bdev_iscsi_set_options", 00:19:54.787 "params": { 00:19:54.787 "timeout_sec": 30 00:19:54.787 } 00:19:54.787 }, 00:19:54.787 { 00:19:54.787 "method": "bdev_nvme_set_options", 00:19:54.787 "params": { 00:19:54.787 "action_on_timeout": "none", 00:19:54.787 "timeout_us": 0, 00:19:54.787 "timeout_admin_us": 0, 00:19:54.787 "keep_alive_timeout_ms": 10000, 00:19:54.787 "arbitration_burst": 0, 00:19:54.787 "low_priority_weight": 0, 00:19:54.787 "medium_priority_weight": 0, 00:19:54.787 "high_priority_weight": 0, 00:19:54.787 "nvme_adminq_poll_period_us": 10000, 00:19:54.787 "nvme_ioq_poll_period_us": 0, 00:19:54.787 "io_queue_requests": 512, 00:19:54.787 "delay_cmd_submit": true, 00:19:54.787 "transport_retry_count": 4, 00:19:54.787 "bdev_retry_count": 3, 00:19:54.788 "transport_ack_timeout": 0, 00:19:54.788 "ctrlr_loss_timeout_sec": 0, 00:19:54.788 "reconnect_delay_sec": 0, 00:19:54.788 "fast_io_fail_timeout_sec": 0, 00:19:54.788 "disable_auto_failback": false, 00:19:54.788 "generate_uuids": false, 00:19:54.788 "transport_tos": 0, 00:19:54.788 "nvme_error_stat": false, 00:19:54.788 "rdma_srq_size": 0, 00:19:54.788 "io_path_stat": false, 00:19:54.788 "allow_accel_sequence": false, 00:19:54.788 "rdma_max_cq_size": 0, 00:19:54.788 "rdma_cm_event_timeout_ms": 0, 00:19:54.788 "dhchap_digests": [ 00:19:54.788 "sha256", 00:19:54.788 "sha384", 00:19:54.788 "sha512" 00:19:54.788 ], 00:19:54.788 "dhchap_dhgroups": [ 00:19:54.788 "null", 00:19:54.788 "ffdhe2048", 00:19:54.788 "ffdhe3072", 00:19:54.788 "ffdhe4096", 00:19:54.788 "ffdhe6144", 00:19:54.788 "ffdhe8192" 00:19:54.788 ], 00:19:54.788 "rdma_umr_per_io": false 00:19:54.788 } 00:19:54.788 }, 00:19:54.788 { 00:19:54.788 "method": "bdev_nvme_attach_controller", 00:19:54.788 "params": { 00:19:54.788 "name": "nvme0", 00:19:54.788 "trtype": "TCP", 00:19:54.788 "adrfam": "IPv4", 00:19:54.788 "traddr": "10.0.0.2", 00:19:54.788 "trsvcid": "4420", 00:19:54.788 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:54.788 "prchk_reftag": false, 00:19:54.788 "prchk_guard": false, 00:19:54.788 "ctrlr_loss_timeout_sec": 0, 00:19:54.788 "reconnect_delay_sec": 0, 00:19:54.788 "fast_io_fail_timeout_sec": 0, 00:19:54.788 "psk": "key0", 00:19:54.788 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:54.788 "hdgst": false, 00:19:54.788 "ddgst": false, 00:19:54.788 "multipath": "multipath" 00:19:54.788 } 00:19:54.788 }, 00:19:54.788 { 00:19:54.788 "method": "bdev_nvme_set_hotplug", 00:19:54.788 "params": { 00:19:54.788 "period_us": 100000, 00:19:54.788 "enable": false 00:19:54.788 } 00:19:54.788 }, 00:19:54.788 { 00:19:54.788 "method": "bdev_enable_histogram", 00:19:54.788 "params": { 00:19:54.788 "name": "nvme0n1", 00:19:54.788 "enable": true 00:19:54.788 } 00:19:54.788 }, 00:19:54.788 { 00:19:54.788 "method": "bdev_wait_for_examine" 00:19:54.788 } 00:19:54.788 ] 00:19:54.788 }, 00:19:54.788 { 00:19:54.788 "subsystem": "nbd", 00:19:54.788 "config": [] 00:19:54.788 } 00:19:54.788 ] 00:19:54.788 }' 00:19:54.788 [2024-10-17 17:41:02.672619] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:19:54.788 [2024-10-17 17:41:02.672671] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73589 ] 00:19:55.048 [2024-10-17 17:41:02.750716] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:55.048 [2024-10-17 17:41:02.780854] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:55.048 [2024-10-17 17:41:02.915604] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:55.617 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:55.617 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:19:55.617 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:19:55.617 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # jq -r '.[].name' 00:19:55.877 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:55.877 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:55.877 Running I/O for 1 seconds... 00:19:57.077 5765.00 IOPS, 22.52 MiB/s 00:19:57.077 Latency(us) 00:19:57.077 [2024-10-17T15:41:04.996Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:57.077 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:57.077 Verification LBA range: start 0x0 length 0x2000 00:19:57.077 nvme0n1 : 1.03 5705.38 22.29 0.00 0.00 22122.30 5707.09 31894.19 00:19:57.077 [2024-10-17T15:41:04.996Z] =================================================================================================================== 00:19:57.077 [2024-10-17T15:41:04.996Z] Total : 5705.38 22.29 0.00 0.00 22122.30 5707.09 31894.19 00:19:57.077 { 00:19:57.077 "results": [ 00:19:57.077 { 00:19:57.077 "job": "nvme0n1", 00:19:57.077 "core_mask": "0x2", 00:19:57.077 "workload": "verify", 00:19:57.077 "status": "finished", 00:19:57.077 "verify_range": { 00:19:57.077 "start": 0, 00:19:57.077 "length": 8192 00:19:57.077 }, 00:19:57.077 "queue_depth": 128, 00:19:57.077 "io_size": 4096, 00:19:57.077 "runtime": 1.03306, 00:19:57.077 "iops": 5705.380132809324, 00:19:57.077 "mibps": 22.286641143786422, 00:19:57.077 "io_failed": 0, 00:19:57.077 "io_timeout": 0, 00:19:57.077 "avg_latency_us": 22122.296319420882, 00:19:57.077 "min_latency_us": 5707.093333333333, 00:19:57.077 "max_latency_us": 31894.18666666667 00:19:57.077 } 00:19:57.077 ], 00:19:57.077 "core_count": 1 00:19:57.077 } 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@282 -- # trap - SIGINT SIGTERM EXIT 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@283 -- # cleanup 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@15 -- # process_shm --id 0 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@808 -- # type=--id 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@809 -- # id=0 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@820 -- # for n in $shm_files 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:19:57.077 nvmf_trace.0 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@823 -- # return 0 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@16 -- # killprocess 73589 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 73589 ']' 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 73589 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73589 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:19:57.077 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73589' 00:19:57.078 killing process with pid 73589 00:19:57.078 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 73589 00:19:57.078 Received shutdown signal, test time was about 1.000000 seconds 00:19:57.078 00:19:57.078 Latency(us) 00:19:57.078 [2024-10-17T15:41:04.997Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:57.078 [2024-10-17T15:41:04.997Z] =================================================================================================================== 00:19:57.078 [2024-10-17T15:41:04.997Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:57.078 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 73589 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@17 -- # nvmftestfini 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@514 -- # nvmfcleanup 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@121 -- # sync 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@124 -- # set +e 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@125 -- # for i in {1..20} 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:19:57.337 rmmod nvme_tcp 00:19:57.337 rmmod nvme_fabrics 00:19:57.337 rmmod nvme_keyring 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@128 -- # set -e 00:19:57.337 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@129 -- # return 0 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@515 -- # '[' -n 73445 ']' 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@516 -- # killprocess 73445 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 73445 ']' 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 73445 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73445 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73445' 00:19:57.338 killing process with pid 73445 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 73445 00:19:57.338 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 73445 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@297 -- # iptr 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@789 -- # iptables-save 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@789 -- # iptables-restore 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@302 -- # remove_spdk_ns 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:57.598 17:41:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@18 -- # rm -f /tmp/tmp.rhGBctHwHP /tmp/tmp.fGOgE4Nk3q /tmp/tmp.wGvnMMapw5 00:19:59.511 00:19:59.511 real 1m27.750s 00:19:59.511 user 2m17.969s 00:19:59.511 sys 0m27.419s 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:59.511 ************************************ 00:19:59.511 END TEST nvmf_tls 00:19:59.511 ************************************ 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@42 -- # run_test nvmf_fips /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:59.511 17:41:07 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:59.771 ************************************ 00:19:59.771 START TEST nvmf_fips 00:19:59.771 ************************************ 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:19:59.771 * Looking for test storage... 00:19:59.771 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1691 -- # lcov --version 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=<' 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=1 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@345 -- # : 1 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=2 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # return 0 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:19:59.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:59.771 --rc genhtml_branch_coverage=1 00:19:59.771 --rc genhtml_function_coverage=1 00:19:59.771 --rc genhtml_legend=1 00:19:59.771 --rc geninfo_all_blocks=1 00:19:59.771 --rc geninfo_unexecuted_blocks=1 00:19:59.771 00:19:59.771 ' 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:19:59.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:59.771 --rc genhtml_branch_coverage=1 00:19:59.771 --rc genhtml_function_coverage=1 00:19:59.771 --rc genhtml_legend=1 00:19:59.771 --rc geninfo_all_blocks=1 00:19:59.771 --rc geninfo_unexecuted_blocks=1 00:19:59.771 00:19:59.771 ' 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:19:59.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:59.771 --rc genhtml_branch_coverage=1 00:19:59.771 --rc genhtml_function_coverage=1 00:19:59.771 --rc genhtml_legend=1 00:19:59.771 --rc geninfo_all_blocks=1 00:19:59.771 --rc geninfo_unexecuted_blocks=1 00:19:59.771 00:19:59.771 ' 00:19:59.771 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:19:59.771 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:59.771 --rc genhtml_branch_coverage=1 00:19:59.771 --rc genhtml_function_coverage=1 00:19:59.771 --rc genhtml_legend=1 00:19:59.771 --rc geninfo_all_blocks=1 00:19:59.771 --rc geninfo_unexecuted_blocks=1 00:19:59.771 00:19:59.772 ' 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # uname -s 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@15 -- # shopt -s extglob 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@5 -- # export PATH 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@51 -- # : 0 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:59.772 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@90 -- # check_openssl_version 00:19:59.772 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@84 -- # local target=3.0.0 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # openssl version 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # awk '{print $2}' 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # ge 3.1.1 3.0.0 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@376 -- # cmp_versions 3.1.1 '>=' 3.0.0 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=>=' 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@348 -- # : 1 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=3 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v++ )) 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 0 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=0 00:20:00.034 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 0 =~ ^[0-9]+$ ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 0 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=0 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # return 0 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # openssl info -modulesdir 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # [[ ! -f /usr/lib64/ossl-modules/fips.so ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # openssl fipsinstall -help 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # warn='This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode' 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@102 -- # [[ This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode == \T\h\i\s\ \c\o\m\m\a\n\d\ \i\s\ \n\o\t\ \e\n\a\b\l\e\d* ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # export callback=build_openssl_config 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # callback=build_openssl_config 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@114 -- # build_openssl_config 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@38 -- # cat 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@58 -- # [[ ! -t 0 ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@59 -- # cat - 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # export OPENSSL_CONF=spdk_fips.conf 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # OPENSSL_CONF=spdk_fips.conf 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # mapfile -t providers 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # openssl list -providers 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # grep name 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # (( 2 != 2 )) 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: openssl base provider != *base* ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: red hat enterprise linux 9 - openssl fips provider != *fips* ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # NOT openssl md5 /dev/fd/62 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@650 -- # local es=0 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@652 -- # valid_exec_arg openssl md5 /dev/fd/62 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # : 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@638 -- # local arg=openssl 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # type -t openssl 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # type -P openssl 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # arg=/usr/bin/openssl 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # [[ -x /usr/bin/openssl ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@653 -- # openssl md5 /dev/fd/62 00:20:00.035 Error setting digest 00:20:00.035 402252B48D7F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:crypto/evp/evp_fetch.c:341:Global default library context, Algorithm (MD5 : 95), Properties () 00:20:00.035 402252B48D7F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:crypto/evp/digest.c:272: 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@653 -- # es=1 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@131 -- # nvmftestinit 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@474 -- # prepare_net_devs 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@436 -- # local -g is_hw=no 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@438 -- # remove_spdk_ns 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@309 -- # xtrace_disable 00:20:00.035 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # pci_devs=() 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # net_devs=() 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # e810=() 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # local -ga e810 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # x722=() 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # local -ga x722 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # mlx=() 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # local -ga mlx 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:20:08.177 Found 0000:31:00.0 (0x8086 - 0x159b) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:20:08.177 Found 0000:31:00.1 (0x8086 - 0x159b) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:20:08.177 Found net devices under 0000:31:00.0: cvl_0_0 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:20:08.177 Found net devices under 0000:31:00.1: cvl_0_1 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # is_hw=yes 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:08.177 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:20:08.178 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:08.178 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.612 ms 00:20:08.178 00:20:08.178 --- 10.0.0.2 ping statistics --- 00:20:08.178 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:08.178 rtt min/avg/max/mdev = 0.612/0.612/0.612/0.000 ms 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:08.178 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:08.178 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.328 ms 00:20:08.178 00:20:08.178 --- 10.0.0.1 ping statistics --- 00:20:08.178 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:08.178 rtt min/avg/max/mdev = 0.328/0.328/0.328/0.000 ms 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@448 -- # return 0 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@132 -- # nvmfappstart -m 0x2 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@724 -- # xtrace_disable 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@507 -- # nvmfpid=78496 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@508 -- # waitforlisten 78496 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@831 -- # '[' -z 78496 ']' 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:08.178 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:08.178 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:20:08.178 [2024-10-17 17:41:15.630791] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:20:08.178 [2024-10-17 17:41:15.630869] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:08.178 [2024-10-17 17:41:15.721205] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:08.178 [2024-10-17 17:41:15.771519] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:08.178 [2024-10-17 17:41:15.771572] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:08.178 [2024-10-17 17:41:15.771582] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:08.178 [2024-10-17 17:41:15.771589] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:08.178 [2024-10-17 17:41:15.771596] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:08.178 [2024-10-17 17:41:15.772388] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # return 0 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@730 -- # xtrace_disable 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@134 -- # trap cleanup EXIT 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@137 -- # key=NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # mktemp -t spdk-psk.XXX 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # key_path=/tmp/spdk-psk.fwF 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@139 -- # echo -n NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@140 -- # chmod 0600 /tmp/spdk-psk.fwF 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@142 -- # setup_nvmf_tgt_conf /tmp/spdk-psk.fwF 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@22 -- # local key=/tmp/spdk-psk.fwF 00:20:08.749 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:08.749 [2024-10-17 17:41:16.658664] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:09.010 [2024-10-17 17:41:16.674644] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:20:09.010 [2024-10-17 17:41:16.675001] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:09.010 malloc0 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@145 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@148 -- # bdevperf_pid=78585 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@149 -- # waitforlisten 78585 /var/tmp/bdevperf.sock 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@146 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@831 -- # '[' -z 78585 ']' 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:09.010 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:09.010 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:20:09.010 [2024-10-17 17:41:16.821173] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:20:09.010 [2024-10-17 17:41:16.821249] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78585 ] 00:20:09.010 [2024-10-17 17:41:16.904744] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:09.271 [2024-10-17 17:41:16.956117] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:09.841 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:09.841 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # return 0 00:20:09.841 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/spdk-psk.fwF 00:20:10.102 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@152 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:20:10.102 [2024-10-17 17:41:17.994023] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:20:10.362 TLSTESTn1 00:20:10.362 17:41:18 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@156 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:20:10.362 Running I/O for 10 seconds... 00:20:12.691 4406.00 IOPS, 17.21 MiB/s [2024-10-17T15:41:21.552Z] 4007.00 IOPS, 15.65 MiB/s [2024-10-17T15:41:22.493Z] 4400.33 IOPS, 17.19 MiB/s [2024-10-17T15:41:23.432Z] 4540.50 IOPS, 17.74 MiB/s [2024-10-17T15:41:24.371Z] 4929.20 IOPS, 19.25 MiB/s [2024-10-17T15:41:25.310Z] 5108.33 IOPS, 19.95 MiB/s [2024-10-17T15:41:26.251Z] 5211.71 IOPS, 20.36 MiB/s [2024-10-17T15:41:27.632Z] 5146.38 IOPS, 20.10 MiB/s [2024-10-17T15:41:28.573Z] 5249.89 IOPS, 20.51 MiB/s [2024-10-17T15:41:28.573Z] 5292.60 IOPS, 20.67 MiB/s 00:20:20.654 Latency(us) 00:20:20.654 [2024-10-17T15:41:28.573Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:20.654 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:20:20.654 Verification LBA range: start 0x0 length 0x2000 00:20:20.654 TLSTESTn1 : 10.01 5297.90 20.69 0.00 0.00 24124.56 5461.33 39540.05 00:20:20.654 [2024-10-17T15:41:28.573Z] =================================================================================================================== 00:20:20.654 [2024-10-17T15:41:28.573Z] Total : 5297.90 20.69 0.00 0.00 24124.56 5461.33 39540.05 00:20:20.654 { 00:20:20.654 "results": [ 00:20:20.654 { 00:20:20.654 "job": "TLSTESTn1", 00:20:20.654 "core_mask": "0x4", 00:20:20.654 "workload": "verify", 00:20:20.654 "status": "finished", 00:20:20.654 "verify_range": { 00:20:20.654 "start": 0, 00:20:20.654 "length": 8192 00:20:20.654 }, 00:20:20.654 "queue_depth": 128, 00:20:20.654 "io_size": 4096, 00:20:20.654 "runtime": 10.013959, 00:20:20.654 "iops": 5297.904654892236, 00:20:20.654 "mibps": 20.694940058172797, 00:20:20.654 "io_failed": 0, 00:20:20.654 "io_timeout": 0, 00:20:20.654 "avg_latency_us": 24124.558566967622, 00:20:20.654 "min_latency_us": 5461.333333333333, 00:20:20.654 "max_latency_us": 39540.05333333334 00:20:20.654 } 00:20:20.654 ], 00:20:20.654 "core_count": 1 00:20:20.654 } 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@1 -- # cleanup 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@15 -- # process_shm --id 0 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@808 -- # type=--id 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@809 -- # id=0 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@820 -- # for n in $shm_files 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:20:20.654 nvmf_trace.0 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@823 -- # return 0 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@16 -- # killprocess 78585 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@950 -- # '[' -z 78585 ']' 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # kill -0 78585 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # uname 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78585 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78585' 00:20:20.654 killing process with pid 78585 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@969 -- # kill 78585 00:20:20.654 Received shutdown signal, test time was about 10.000000 seconds 00:20:20.654 00:20:20.654 Latency(us) 00:20:20.654 [2024-10-17T15:41:28.573Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:20.654 [2024-10-17T15:41:28.573Z] =================================================================================================================== 00:20:20.654 [2024-10-17T15:41:28.573Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@974 -- # wait 78585 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@17 -- # nvmftestfini 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@514 -- # nvmfcleanup 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@121 -- # sync 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@124 -- # set +e 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@125 -- # for i in {1..20} 00:20:20.654 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:20:20.654 rmmod nvme_tcp 00:20:20.654 rmmod nvme_fabrics 00:20:20.654 rmmod nvme_keyring 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@128 -- # set -e 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@129 -- # return 0 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@515 -- # '[' -n 78496 ']' 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@516 -- # killprocess 78496 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@950 -- # '[' -z 78496 ']' 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # kill -0 78496 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # uname 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78496 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:20:20.914 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78496' 00:20:20.915 killing process with pid 78496 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@969 -- # kill 78496 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@974 -- # wait 78496 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@297 -- # iptr 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@789 -- # iptables-save 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@789 -- # iptables-restore 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@302 -- # remove_spdk_ns 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:20.915 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:23.460 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:20:23.460 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@18 -- # rm -f /tmp/spdk-psk.fwF 00:20:23.460 00:20:23.460 real 0m23.387s 00:20:23.460 user 0m24.877s 00:20:23.460 sys 0m9.889s 00:20:23.460 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:23.461 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:20:23.461 ************************************ 00:20:23.461 END TEST nvmf_fips 00:20:23.461 ************************************ 00:20:23.461 17:41:30 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@43 -- # run_test nvmf_control_msg_list /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:20:23.461 17:41:30 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:20:23.461 17:41:30 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:23.461 17:41:30 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:23.461 ************************************ 00:20:23.461 START TEST nvmf_control_msg_list 00:20:23.461 ************************************ 00:20:23.461 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:20:23.461 * Looking for test storage... 00:20:23.461 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1691 -- # lcov --version 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # IFS=.-: 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # read -ra ver1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # IFS=.-: 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # read -ra ver2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@338 -- # local 'op=<' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@340 -- # ver1_l=2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@341 -- # ver2_l=1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@344 -- # case "$op" in 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@345 -- # : 1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # decimal 1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # ver1[v]=1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # decimal 2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # ver2[v]=2 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # return 0 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:20:23.461 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:23.461 --rc genhtml_branch_coverage=1 00:20:23.461 --rc genhtml_function_coverage=1 00:20:23.461 --rc genhtml_legend=1 00:20:23.461 --rc geninfo_all_blocks=1 00:20:23.461 --rc geninfo_unexecuted_blocks=1 00:20:23.461 00:20:23.461 ' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:20:23.461 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:23.461 --rc genhtml_branch_coverage=1 00:20:23.461 --rc genhtml_function_coverage=1 00:20:23.461 --rc genhtml_legend=1 00:20:23.461 --rc geninfo_all_blocks=1 00:20:23.461 --rc geninfo_unexecuted_blocks=1 00:20:23.461 00:20:23.461 ' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:20:23.461 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:23.461 --rc genhtml_branch_coverage=1 00:20:23.461 --rc genhtml_function_coverage=1 00:20:23.461 --rc genhtml_legend=1 00:20:23.461 --rc geninfo_all_blocks=1 00:20:23.461 --rc geninfo_unexecuted_blocks=1 00:20:23.461 00:20:23.461 ' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:20:23.461 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:23.461 --rc genhtml_branch_coverage=1 00:20:23.461 --rc genhtml_function_coverage=1 00:20:23.461 --rc genhtml_legend=1 00:20:23.461 --rc geninfo_all_blocks=1 00:20:23.461 --rc geninfo_unexecuted_blocks=1 00:20:23.461 00:20:23.461 ' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # uname -s 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@15 -- # shopt -s extglob 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@5 -- # export PATH 00:20:23.461 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@51 -- # : 0 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:23.462 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@12 -- # nvmftestinit 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@474 -- # prepare_net_devs 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@436 -- # local -g is_hw=no 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@438 -- # remove_spdk_ns 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@309 -- # xtrace_disable 00:20:23.462 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # pci_devs=() 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # net_devs=() 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # e810=() 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # local -ga e810 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # x722=() 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # local -ga x722 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # mlx=() 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # local -ga mlx 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:20:31.606 Found 0000:31:00.0 (0x8086 - 0x159b) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:20:31.606 Found 0000:31:00.1 (0x8086 - 0x159b) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:31.606 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:20:31.607 Found net devices under 0000:31:00.0: cvl_0_0 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:20:31.607 Found net devices under 0000:31:00.1: cvl_0_1 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # is_hw=yes 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:20:31.607 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:31.607 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.641 ms 00:20:31.607 00:20:31.607 --- 10.0.0.2 ping statistics --- 00:20:31.607 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:31.607 rtt min/avg/max/mdev = 0.641/0.641/0.641/0.000 ms 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:31.607 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:31.607 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.356 ms 00:20:31.607 00:20:31.607 --- 10.0.0.1 ping statistics --- 00:20:31.607 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:31.607 rtt min/avg/max/mdev = 0.356/0.356/0.356/0.000 ms 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@448 -- # return 0 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@13 -- # nvmfappstart 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@724 -- # xtrace_disable 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@507 -- # nvmfpid=85261 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@508 -- # waitforlisten 85261 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@831 -- # '[' -z 85261 ']' 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:31.607 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:31.607 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:31.608 [2024-10-17 17:41:38.916495] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:20:31.608 [2024-10-17 17:41:38.916562] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:31.608 [2024-10-17 17:41:39.008747] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:31.608 [2024-10-17 17:41:39.060016] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:31.608 [2024-10-17 17:41:39.060069] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:31.608 [2024-10-17 17:41:39.060078] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:31.608 [2024-10-17 17:41:39.060085] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:31.608 [2024-10-17 17:41:39.060091] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:31.608 [2024-10-17 17:41:39.060941] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@864 -- # return 0 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@730 -- # xtrace_disable 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@19 -- # rpc_cmd nvmf_create_transport '-t tcp -o' --in-capsule-data-size 768 --control-msg-num 1 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:31.869 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:32.131 [2024-10-17 17:41:39.789298] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@21 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:32.131 Malloc0 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:32.131 [2024-10-17 17:41:39.843981] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@27 -- # perf_pid1=85299 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x2 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@29 -- # perf_pid2=85301 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x4 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@31 -- # perf_pid3=85303 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@33 -- # wait 85299 00:20:32.131 17:41:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x8 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:20:32.131 [2024-10-17 17:41:39.914582] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:20:32.131 [2024-10-17 17:41:39.914845] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:20:32.131 [2024-10-17 17:41:39.924604] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:20:33.517 Initializing NVMe Controllers 00:20:33.517 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:20:33.517 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 3 00:20:33.517 Initialization complete. Launching workers. 00:20:33.517 ======================================================== 00:20:33.517 Latency(us) 00:20:33.517 Device Information : IOPS MiB/s Average min max 00:20:33.517 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 3: 1598.00 6.24 625.48 182.75 1300.84 00:20:33.517 ======================================================== 00:20:33.517 Total : 1598.00 6.24 625.48 182.75 1300.84 00:20:33.517 00:20:33.517 Initializing NVMe Controllers 00:20:33.517 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:20:33.517 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 2 00:20:33.517 Initialization complete. Launching workers. 00:20:33.517 ======================================================== 00:20:33.517 Latency(us) 00:20:33.517 Device Information : IOPS MiB/s Average min max 00:20:33.517 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 2: 1685.00 6.58 593.20 146.43 954.41 00:20:33.517 ======================================================== 00:20:33.517 Total : 1685.00 6.58 593.20 146.43 954.41 00:20:33.517 00:20:33.517 Initializing NVMe Controllers 00:20:33.517 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:20:33.517 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 1 00:20:33.517 Initialization complete. Launching workers. 00:20:33.517 ======================================================== 00:20:33.517 Latency(us) 00:20:33.517 Device Information : IOPS MiB/s Average min max 00:20:33.517 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 1: 1806.00 7.05 553.65 154.05 1418.09 00:20:33.517 ======================================================== 00:20:33.517 Total : 1806.00 7.05 553.65 154.05 1418.09 00:20:33.517 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@34 -- # wait 85301 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@35 -- # wait 85303 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@38 -- # nvmftestfini 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@514 -- # nvmfcleanup 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@121 -- # sync 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@124 -- # set +e 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@125 -- # for i in {1..20} 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:20:33.517 rmmod nvme_tcp 00:20:33.517 rmmod nvme_fabrics 00:20:33.517 rmmod nvme_keyring 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@128 -- # set -e 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@129 -- # return 0 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@515 -- # '[' -n 85261 ']' 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@516 -- # killprocess 85261 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@950 -- # '[' -z 85261 ']' 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@954 -- # kill -0 85261 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@955 -- # uname 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85261 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85261' 00:20:33.517 killing process with pid 85261 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@969 -- # kill 85261 00:20:33.517 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@974 -- # wait 85261 00:20:33.778 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@297 -- # iptr 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@789 -- # iptables-save 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@789 -- # iptables-restore 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@302 -- # remove_spdk_ns 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:33.779 17:41:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:35.691 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:20:35.691 00:20:35.691 real 0m12.624s 00:20:35.691 user 0m8.102s 00:20:35.691 sys 0m6.803s 00:20:35.691 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:35.691 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:20:35.691 ************************************ 00:20:35.691 END TEST nvmf_control_msg_list 00:20:35.691 ************************************ 00:20:35.691 17:41:43 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@44 -- # run_test nvmf_wait_for_buf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:20:35.691 17:41:43 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:20:35.691 17:41:43 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:35.691 17:41:43 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:35.952 ************************************ 00:20:35.952 START TEST nvmf_wait_for_buf 00:20:35.952 ************************************ 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:20:35.952 * Looking for test storage... 00:20:35.952 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1691 -- # lcov --version 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # IFS=.-: 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # read -ra ver1 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # IFS=.-: 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # read -ra ver2 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@338 -- # local 'op=<' 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@340 -- # ver1_l=2 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@341 -- # ver2_l=1 00:20:35.952 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@344 -- # case "$op" in 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@345 -- # : 1 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # decimal 1 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=1 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 1 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # ver1[v]=1 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # decimal 2 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=2 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 2 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # ver2[v]=2 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # return 0 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:20:35.953 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:35.953 --rc genhtml_branch_coverage=1 00:20:35.953 --rc genhtml_function_coverage=1 00:20:35.953 --rc genhtml_legend=1 00:20:35.953 --rc geninfo_all_blocks=1 00:20:35.953 --rc geninfo_unexecuted_blocks=1 00:20:35.953 00:20:35.953 ' 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:20:35.953 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:35.953 --rc genhtml_branch_coverage=1 00:20:35.953 --rc genhtml_function_coverage=1 00:20:35.953 --rc genhtml_legend=1 00:20:35.953 --rc geninfo_all_blocks=1 00:20:35.953 --rc geninfo_unexecuted_blocks=1 00:20:35.953 00:20:35.953 ' 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:20:35.953 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:35.953 --rc genhtml_branch_coverage=1 00:20:35.953 --rc genhtml_function_coverage=1 00:20:35.953 --rc genhtml_legend=1 00:20:35.953 --rc geninfo_all_blocks=1 00:20:35.953 --rc geninfo_unexecuted_blocks=1 00:20:35.953 00:20:35.953 ' 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:20:35.953 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:35.953 --rc genhtml_branch_coverage=1 00:20:35.953 --rc genhtml_function_coverage=1 00:20:35.953 --rc genhtml_legend=1 00:20:35.953 --rc geninfo_all_blocks=1 00:20:35.953 --rc geninfo_unexecuted_blocks=1 00:20:35.953 00:20:35.953 ' 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # uname -s 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:35.953 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@15 -- # shopt -s extglob 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:36.305 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@5 -- # export PATH 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@51 -- # : 0 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:36.306 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@12 -- # nvmftestinit 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@474 -- # prepare_net_devs 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@436 -- # local -g is_hw=no 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@438 -- # remove_spdk_ns 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@309 -- # xtrace_disable 00:20:36.306 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # pci_devs=() 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # net_devs=() 00:20:44.517 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # e810=() 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # local -ga e810 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # x722=() 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # local -ga x722 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # mlx=() 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # local -ga mlx 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:20:44.518 Found 0000:31:00.0 (0x8086 - 0x159b) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:20:44.518 Found 0000:31:00.1 (0x8086 - 0x159b) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:20:44.518 Found net devices under 0000:31:00.0: cvl_0_0 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:20:44.518 Found net devices under 0000:31:00.1: cvl_0_1 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # is_hw=yes 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:20:44.518 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:44.518 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.698 ms 00:20:44.518 00:20:44.518 --- 10.0.0.2 ping statistics --- 00:20:44.518 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:44.518 rtt min/avg/max/mdev = 0.698/0.698/0.698/0.000 ms 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:44.518 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:44.518 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.237 ms 00:20:44.518 00:20:44.518 --- 10.0.0.1 ping statistics --- 00:20:44.518 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:44.518 rtt min/avg/max/mdev = 0.237/0.237/0.237/0.000 ms 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@448 -- # return 0 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@13 -- # nvmfappstart --wait-for-rpc 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@724 -- # xtrace_disable 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@507 -- # nvmfpid=89985 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@508 -- # waitforlisten 89985 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@831 -- # '[' -z 89985 ']' 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:44.518 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:44.518 17:41:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.518 [2024-10-17 17:41:51.664873] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:20:44.518 [2024-10-17 17:41:51.664937] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:44.518 [2024-10-17 17:41:51.753419] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:44.518 [2024-10-17 17:41:51.805048] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:44.518 [2024-10-17 17:41:51.805098] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:44.518 [2024-10-17 17:41:51.805106] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:44.518 [2024-10-17 17:41:51.805113] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:44.518 [2024-10-17 17:41:51.805120] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:44.518 [2024-10-17 17:41:51.805910] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@864 -- # return 0 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@730 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@19 -- # rpc_cmd accel_set_options --small-cache-size 0 --large-cache-size 0 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@20 -- # rpc_cmd iobuf_set_options --small-pool-count 154 --small_bufsize=8192 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@21 -- # rpc_cmd framework_start_init 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@22 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 Malloc0 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@23 -- # rpc_cmd nvmf_create_transport '-t tcp -o' -u 8192 -n 24 -b 24 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 [2024-10-17 17:41:52.630626] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a -s SPDK00000000000001 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:44.779 [2024-10-17 17:41:52.654918] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.779 17:41:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 4 -o 131072 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:20:45.041 [2024-10-17 17:41:52.740807] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:20:46.522 Initializing NVMe Controllers 00:20:46.522 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:20:46.522 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 0 00:20:46.522 Initialization complete. Launching workers. 00:20:46.522 ======================================================== 00:20:46.522 Latency(us) 00:20:46.522 Device Information : IOPS MiB/s Average min max 00:20:46.522 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 0: 24.98 3.12 167749.83 47866.14 199531.36 00:20:46.522 ======================================================== 00:20:46.522 Total : 24.98 3.12 167749.83 47866.14 199531.36 00:20:46.522 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # rpc_cmd iobuf_get_stats 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # jq -r '.[] | select(.module == "nvmf_TCP") | .small_pool.retry' 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # retry_count=374 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@33 -- # [[ 374 -eq 0 ]] 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@38 -- # nvmftestfini 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@514 -- # nvmfcleanup 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@121 -- # sync 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@124 -- # set +e 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@125 -- # for i in {1..20} 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:20:46.522 rmmod nvme_tcp 00:20:46.522 rmmod nvme_fabrics 00:20:46.522 rmmod nvme_keyring 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@128 -- # set -e 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@129 -- # return 0 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@515 -- # '[' -n 89985 ']' 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@516 -- # killprocess 89985 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@950 -- # '[' -z 89985 ']' 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@954 -- # kill -0 89985 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@955 -- # uname 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89985 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89985' 00:20:46.522 killing process with pid 89985 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@969 -- # kill 89985 00:20:46.522 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@974 -- # wait 89985 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@297 -- # iptr 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@789 -- # iptables-save 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@789 -- # iptables-restore 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:46.784 17:41:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:48.700 17:41:56 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:20:48.700 00:20:48.700 real 0m12.970s 00:20:48.700 user 0m5.243s 00:20:48.700 sys 0m6.289s 00:20:48.700 17:41:56 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:48.700 17:41:56 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:20:48.700 ************************************ 00:20:48.700 END TEST nvmf_wait_for_buf 00:20:48.700 ************************************ 00:20:48.962 17:41:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@47 -- # '[' 0 -eq 1 ']' 00:20:48.962 17:41:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@53 -- # [[ phy == phy ]] 00:20:48.962 17:41:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@54 -- # '[' tcp = tcp ']' 00:20:48.962 17:41:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@55 -- # gather_supported_nvmf_pci_devs 00:20:48.962 17:41:56 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@309 -- # xtrace_disable 00:20:48.962 17:41:56 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # pci_devs=() 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # net_devs=() 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # e810=() 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # local -ga e810 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # x722=() 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # local -ga x722 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # mlx=() 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # local -ga mlx 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:20:57.107 Found 0000:31:00.0 (0x8086 - 0x159b) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:20:57.107 Found 0000:31:00.1 (0x8086 - 0x159b) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:20:57.107 Found net devices under 0000:31:00.0: cvl_0_0 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:20:57.107 Found net devices under 0000:31:00.1: cvl_0_1 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@56 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@57 -- # (( 2 > 0 )) 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@58 -- # run_test nvmf_perf_adq /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:57.107 17:42:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:57.107 ************************************ 00:20:57.107 START TEST nvmf_perf_adq 00:20:57.107 ************************************ 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:20:57.107 * Looking for test storage... 00:20:57.107 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1691 -- # lcov --version 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:57.107 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # IFS=.-: 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # read -ra ver1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # IFS=.-: 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # read -ra ver2 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@338 -- # local 'op=<' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@340 -- # ver1_l=2 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@341 -- # ver2_l=1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@344 -- # case "$op" in 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@345 -- # : 1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # decimal 1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # ver1[v]=1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # decimal 2 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=2 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 2 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # ver2[v]=2 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # return 0 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:20:57.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:57.108 --rc genhtml_branch_coverage=1 00:20:57.108 --rc genhtml_function_coverage=1 00:20:57.108 --rc genhtml_legend=1 00:20:57.108 --rc geninfo_all_blocks=1 00:20:57.108 --rc geninfo_unexecuted_blocks=1 00:20:57.108 00:20:57.108 ' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:20:57.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:57.108 --rc genhtml_branch_coverage=1 00:20:57.108 --rc genhtml_function_coverage=1 00:20:57.108 --rc genhtml_legend=1 00:20:57.108 --rc geninfo_all_blocks=1 00:20:57.108 --rc geninfo_unexecuted_blocks=1 00:20:57.108 00:20:57.108 ' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:20:57.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:57.108 --rc genhtml_branch_coverage=1 00:20:57.108 --rc genhtml_function_coverage=1 00:20:57.108 --rc genhtml_legend=1 00:20:57.108 --rc geninfo_all_blocks=1 00:20:57.108 --rc geninfo_unexecuted_blocks=1 00:20:57.108 00:20:57.108 ' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:20:57.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:57.108 --rc genhtml_branch_coverage=1 00:20:57.108 --rc genhtml_function_coverage=1 00:20:57.108 --rc genhtml_legend=1 00:20:57.108 --rc geninfo_all_blocks=1 00:20:57.108 --rc geninfo_unexecuted_blocks=1 00:20:57.108 00:20:57.108 ' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # uname -s 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@15 -- # shopt -s extglob 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@5 -- # export PATH 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@51 -- # : 0 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:57.108 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@11 -- # gather_supported_nvmf_pci_devs 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:20:57.108 17:42:04 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:21:03.697 Found 0000:31:00.0 (0x8086 - 0x159b) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:21:03.697 Found 0000:31:00.1 (0x8086 - 0x159b) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:21:03.697 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:21:03.698 Found net devices under 0000:31:00.0: cvl_0_0 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:21:03.698 Found net devices under 0000:31:00.1: cvl_0_1 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@12 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@13 -- # (( 2 == 0 )) 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@18 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@68 -- # adq_reload_driver 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:21:03.698 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:21:05.616 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:21:08.172 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@76 -- # nvmftestinit 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # prepare_net_devs 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@436 -- # local -g is_hw=no 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # remove_spdk_ns 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:21:13.476 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:21:13.477 Found 0000:31:00.0 (0x8086 - 0x159b) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:21:13.477 Found 0000:31:00.1 (0x8086 - 0x159b) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:21:13.477 Found net devices under 0000:31:00.0: cvl_0_0 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:21:13.477 Found net devices under 0000:31:00.1: cvl_0_1 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # is_hw=yes 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:21:13.477 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:13.477 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.611 ms 00:21:13.477 00:21:13.477 --- 10.0.0.2 ping statistics --- 00:21:13.477 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:13.477 rtt min/avg/max/mdev = 0.611/0.611/0.611/0.000 ms 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:13.477 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:13.477 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.295 ms 00:21:13.477 00:21:13.477 --- 10.0.0.1 ping statistics --- 00:21:13.477 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:13.477 rtt min/avg/max/mdev = 0.295/0.295/0.295/0.000 ms 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@448 -- # return 0 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@77 -- # nvmfappstart -m 0xF --wait-for-rpc 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@724 -- # xtrace_disable 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # nvmfpid=100869 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # waitforlisten 100869 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:21:13.477 17:42:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@831 -- # '[' -z 100869 ']' 00:21:13.477 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:13.477 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:13.477 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:13.477 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:13.477 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:13.477 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:13.477 [2024-10-17 17:42:21.057925] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:21:13.477 [2024-10-17 17:42:21.057998] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:13.477 [2024-10-17 17:42:21.149901] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:13.477 [2024-10-17 17:42:21.203965] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:13.477 [2024-10-17 17:42:21.204024] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:13.477 [2024-10-17 17:42:21.204037] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:13.477 [2024-10-17 17:42:21.204045] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:13.477 [2024-10-17 17:42:21.204051] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:13.477 [2024-10-17 17:42:21.206502] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:13.477 [2024-10-17 17:42:21.206653] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:21:13.477 [2024-10-17 17:42:21.206812] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:21:13.477 [2024-10-17 17:42:21.206980] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # return 0 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@730 -- # xtrace_disable 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@78 -- # adq_configure_nvmf_target 0 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.050 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 0 --enable-zerocopy-send-server -i posix 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.311 17:42:21 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 0 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.311 [2024-10-17 17:42:22.079274] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.311 Malloc1 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:14.311 [2024-10-17 17:42:22.153963] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@82 -- # perfpid=101219 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@83 -- # sleep 2 00:21:14.311 17:42:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:21:16.854 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # rpc_cmd nvmf_get_stats 00:21:16.854 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.854 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:16.854 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.855 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # nvmf_stats='{ 00:21:16.855 "tick_rate": 2400000000, 00:21:16.855 "poll_groups": [ 00:21:16.855 { 00:21:16.855 "name": "nvmf_tgt_poll_group_000", 00:21:16.855 "admin_qpairs": 1, 00:21:16.855 "io_qpairs": 1, 00:21:16.855 "current_admin_qpairs": 1, 00:21:16.855 "current_io_qpairs": 1, 00:21:16.855 "pending_bdev_io": 0, 00:21:16.855 "completed_nvme_io": 17861, 00:21:16.855 "transports": [ 00:21:16.855 { 00:21:16.855 "trtype": "TCP" 00:21:16.855 } 00:21:16.855 ] 00:21:16.855 }, 00:21:16.855 { 00:21:16.855 "name": "nvmf_tgt_poll_group_001", 00:21:16.855 "admin_qpairs": 0, 00:21:16.855 "io_qpairs": 1, 00:21:16.855 "current_admin_qpairs": 0, 00:21:16.855 "current_io_qpairs": 1, 00:21:16.855 "pending_bdev_io": 0, 00:21:16.855 "completed_nvme_io": 18995, 00:21:16.855 "transports": [ 00:21:16.855 { 00:21:16.855 "trtype": "TCP" 00:21:16.855 } 00:21:16.855 ] 00:21:16.855 }, 00:21:16.855 { 00:21:16.855 "name": "nvmf_tgt_poll_group_002", 00:21:16.855 "admin_qpairs": 0, 00:21:16.855 "io_qpairs": 1, 00:21:16.855 "current_admin_qpairs": 0, 00:21:16.855 "current_io_qpairs": 1, 00:21:16.855 "pending_bdev_io": 0, 00:21:16.855 "completed_nvme_io": 18774, 00:21:16.855 "transports": [ 00:21:16.855 { 00:21:16.855 "trtype": "TCP" 00:21:16.855 } 00:21:16.855 ] 00:21:16.855 }, 00:21:16.855 { 00:21:16.855 "name": "nvmf_tgt_poll_group_003", 00:21:16.855 "admin_qpairs": 0, 00:21:16.855 "io_qpairs": 1, 00:21:16.855 "current_admin_qpairs": 0, 00:21:16.855 "current_io_qpairs": 1, 00:21:16.855 "pending_bdev_io": 0, 00:21:16.855 "completed_nvme_io": 16907, 00:21:16.855 "transports": [ 00:21:16.855 { 00:21:16.855 "trtype": "TCP" 00:21:16.855 } 00:21:16.855 ] 00:21:16.855 } 00:21:16.855 ] 00:21:16.855 }' 00:21:16.855 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 1) | length' 00:21:16.855 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # wc -l 00:21:16.855 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # count=4 00:21:16.855 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@87 -- # [[ 4 -ne 4 ]] 00:21:16.855 17:42:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@91 -- # wait 101219 00:21:24.990 Initializing NVMe Controllers 00:21:24.990 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:24.990 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:21:24.990 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:21:24.990 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:21:24.990 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:21:24.990 Initialization complete. Launching workers. 00:21:24.990 ======================================================== 00:21:24.990 Latency(us) 00:21:24.990 Device Information : IOPS MiB/s Average min max 00:21:24.990 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 13306.00 51.98 4824.82 1300.58 45159.12 00:21:24.990 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 13504.40 52.75 4739.04 1205.67 14023.48 00:21:24.990 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 12631.40 49.34 5066.24 1388.03 13854.53 00:21:24.990 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 13229.20 51.68 4837.01 937.77 12751.76 00:21:24.990 ======================================================== 00:21:24.990 Total : 52671.00 205.75 4863.78 937.77 45159.12 00:21:24.990 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@92 -- # nvmftestfini 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@514 -- # nvmfcleanup 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:21:24.990 rmmod nvme_tcp 00:21:24.990 rmmod nvme_fabrics 00:21:24.990 rmmod nvme_keyring 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@515 -- # '[' -n 100869 ']' 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # killprocess 100869 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@950 -- # '[' -z 100869 ']' 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # kill -0 100869 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # uname 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 100869 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@968 -- # echo 'killing process with pid 100869' 00:21:24.990 killing process with pid 100869 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@969 -- # kill 100869 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@974 -- # wait 100869 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-save 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-restore 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:24.990 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:26.901 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:21:26.901 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@94 -- # adq_reload_driver 00:21:26.901 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:21:26.901 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:21:28.814 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:21:30.726 17:42:38 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@97 -- # nvmftestinit 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # prepare_net_devs 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@436 -- # local -g is_hw=no 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # remove_spdk_ns 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:21:36.017 Found 0000:31:00.0 (0x8086 - 0x159b) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:21:36.017 Found 0000:31:00.1 (0x8086 - 0x159b) 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:36.017 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:21:36.018 Found net devices under 0000:31:00.0: cvl_0_0 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:21:36.018 Found net devices under 0000:31:00.1: cvl_0_1 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # is_hw=yes 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:21:36.018 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:36.018 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.647 ms 00:21:36.018 00:21:36.018 --- 10.0.0.2 ping statistics --- 00:21:36.018 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:36.018 rtt min/avg/max/mdev = 0.647/0.647/0.647/0.000 ms 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:36.018 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:36.018 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.310 ms 00:21:36.018 00:21:36.018 --- 10.0.0.1 ping statistics --- 00:21:36.018 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:36.018 rtt min/avg/max/mdev = 0.310/0.310/0.310/0.000 ms 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@448 -- # return 0 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@98 -- # adq_configure_driver 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@22 -- # ip netns exec cvl_0_0_ns_spdk ethtool --offload cvl_0_0 hw-tc-offload on 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@24 -- # ip netns exec cvl_0_0_ns_spdk ethtool --set-priv-flags cvl_0_0 channel-pkt-inspect-optimize off 00:21:36.018 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@26 -- # sysctl -w net.core.busy_poll=1 00:21:36.018 net.core.busy_poll = 1 00:21:36.279 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@27 -- # sysctl -w net.core.busy_read=1 00:21:36.279 net.core.busy_read = 1 00:21:36.279 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@29 -- # tc=/usr/sbin/tc 00:21:36.279 17:42:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@31 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 root mqprio num_tc 2 map 0 1 queues 2@0 2@2 hw 1 mode channel 00:21:36.279 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 ingress 00:21:36.279 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@35 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc filter add dev cvl_0_0 protocol ip parent ffff: prio 1 flower dst_ip 10.0.0.2/32 ip_proto tcp dst_port 4420 skip_sw hw_tc 1 00:21:36.279 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@38 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/nvmf/set_xps_rxqs cvl_0_0 00:21:36.279 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@99 -- # nvmfappstart -m 0xF --wait-for-rpc 00:21:36.279 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:21:36.279 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@724 -- # xtrace_disable 00:21:36.279 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # nvmfpid=105679 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # waitforlisten 105679 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@831 -- # '[' -z 105679 ']' 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:36.541 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:36.541 17:42:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:36.541 [2024-10-17 17:42:44.259431] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:21:36.541 [2024-10-17 17:42:44.259495] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:36.541 [2024-10-17 17:42:44.351334] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:36.541 [2024-10-17 17:42:44.404684] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:36.541 [2024-10-17 17:42:44.404744] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:36.541 [2024-10-17 17:42:44.404753] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:36.541 [2024-10-17 17:42:44.404762] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:36.541 [2024-10-17 17:42:44.404768] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:36.541 [2024-10-17 17:42:44.406802] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:36.541 [2024-10-17 17:42:44.406877] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:21:36.541 [2024-10-17 17:42:44.407036] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:36.541 [2024-10-17 17:42:44.407035] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # return 0 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@730 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@100 -- # adq_configure_nvmf_target 1 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 1 --enable-zerocopy-send-server -i posix 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 1 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 [2024-10-17 17:42:45.279227] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 Malloc1 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:37.487 [2024-10-17 17:42:45.354024] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@104 -- # perfpid=106033 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@105 -- # sleep 2 00:21:37.487 17:42:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # rpc_cmd nvmf_get_stats 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # nvmf_stats='{ 00:21:40.035 "tick_rate": 2400000000, 00:21:40.035 "poll_groups": [ 00:21:40.035 { 00:21:40.035 "name": "nvmf_tgt_poll_group_000", 00:21:40.035 "admin_qpairs": 1, 00:21:40.035 "io_qpairs": 3, 00:21:40.035 "current_admin_qpairs": 1, 00:21:40.035 "current_io_qpairs": 3, 00:21:40.035 "pending_bdev_io": 0, 00:21:40.035 "completed_nvme_io": 25730, 00:21:40.035 "transports": [ 00:21:40.035 { 00:21:40.035 "trtype": "TCP" 00:21:40.035 } 00:21:40.035 ] 00:21:40.035 }, 00:21:40.035 { 00:21:40.035 "name": "nvmf_tgt_poll_group_001", 00:21:40.035 "admin_qpairs": 0, 00:21:40.035 "io_qpairs": 1, 00:21:40.035 "current_admin_qpairs": 0, 00:21:40.035 "current_io_qpairs": 1, 00:21:40.035 "pending_bdev_io": 0, 00:21:40.035 "completed_nvme_io": 22954, 00:21:40.035 "transports": [ 00:21:40.035 { 00:21:40.035 "trtype": "TCP" 00:21:40.035 } 00:21:40.035 ] 00:21:40.035 }, 00:21:40.035 { 00:21:40.035 "name": "nvmf_tgt_poll_group_002", 00:21:40.035 "admin_qpairs": 0, 00:21:40.035 "io_qpairs": 0, 00:21:40.035 "current_admin_qpairs": 0, 00:21:40.035 "current_io_qpairs": 0, 00:21:40.035 "pending_bdev_io": 0, 00:21:40.035 "completed_nvme_io": 0, 00:21:40.035 "transports": [ 00:21:40.035 { 00:21:40.035 "trtype": "TCP" 00:21:40.035 } 00:21:40.035 ] 00:21:40.035 }, 00:21:40.035 { 00:21:40.035 "name": "nvmf_tgt_poll_group_003", 00:21:40.035 "admin_qpairs": 0, 00:21:40.035 "io_qpairs": 0, 00:21:40.035 "current_admin_qpairs": 0, 00:21:40.035 "current_io_qpairs": 0, 00:21:40.035 "pending_bdev_io": 0, 00:21:40.035 "completed_nvme_io": 0, 00:21:40.035 "transports": [ 00:21:40.035 { 00:21:40.035 "trtype": "TCP" 00:21:40.035 } 00:21:40.035 ] 00:21:40.035 } 00:21:40.035 ] 00:21:40.035 }' 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 0) | length' 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # wc -l 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # count=2 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@109 -- # [[ 2 -lt 2 ]] 00:21:40.035 17:42:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@114 -- # wait 106033 00:21:48.178 Initializing NVMe Controllers 00:21:48.178 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:48.178 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:21:48.178 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:21:48.178 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:21:48.178 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:21:48.178 Initialization complete. Launching workers. 00:21:48.178 ======================================================== 00:21:48.178 Latency(us) 00:21:48.178 Device Information : IOPS MiB/s Average min max 00:21:48.178 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 17609.39 68.79 3634.47 992.14 44373.46 00:21:48.178 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 7624.30 29.78 8393.27 1367.87 61793.53 00:21:48.178 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 6214.00 24.27 10298.86 1187.71 60850.95 00:21:48.178 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 5893.20 23.02 10887.38 1304.15 62550.98 00:21:48.178 ======================================================== 00:21:48.178 Total : 37340.89 145.86 6859.83 992.14 62550.98 00:21:48.178 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@115 -- # nvmftestfini 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@514 -- # nvmfcleanup 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:21:48.178 rmmod nvme_tcp 00:21:48.178 rmmod nvme_fabrics 00:21:48.178 rmmod nvme_keyring 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@515 -- # '[' -n 105679 ']' 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # killprocess 105679 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@950 -- # '[' -z 105679 ']' 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # kill -0 105679 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # uname 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 105679 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@968 -- # echo 'killing process with pid 105679' 00:21:48.178 killing process with pid 105679 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@969 -- # kill 105679 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@974 -- # wait 105679 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-save 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-restore 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:21:48.178 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:21:48.179 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:48.179 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:48.179 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@117 -- # trap - SIGINT SIGTERM EXIT 00:21:50.090 00:21:50.090 real 0m53.859s 00:21:50.090 user 2m50.281s 00:21:50.090 sys 0m11.692s 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:21:50.090 ************************************ 00:21:50.090 END TEST nvmf_perf_adq 00:21:50.090 ************************************ 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@65 -- # run_test nvmf_shutdown /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:21:50.090 ************************************ 00:21:50.090 START TEST nvmf_shutdown 00:21:50.090 ************************************ 00:21:50.090 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:21:50.351 * Looking for test storage... 00:21:50.351 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1691 -- # lcov --version 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@333 -- # local ver1 ver1_l 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@334 -- # local ver2 ver2_l 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # IFS=.-: 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # read -ra ver1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # IFS=.-: 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # read -ra ver2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@338 -- # local 'op=<' 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@340 -- # ver1_l=2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@341 -- # ver2_l=1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@344 -- # case "$op" in 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@345 -- # : 1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v = 0 )) 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # decimal 1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # ver1[v]=1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # decimal 2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # ver2[v]=2 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # return 0 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:21:50.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:50.351 --rc genhtml_branch_coverage=1 00:21:50.351 --rc genhtml_function_coverage=1 00:21:50.351 --rc genhtml_legend=1 00:21:50.351 --rc geninfo_all_blocks=1 00:21:50.351 --rc geninfo_unexecuted_blocks=1 00:21:50.351 00:21:50.351 ' 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:21:50.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:50.351 --rc genhtml_branch_coverage=1 00:21:50.351 --rc genhtml_function_coverage=1 00:21:50.351 --rc genhtml_legend=1 00:21:50.351 --rc geninfo_all_blocks=1 00:21:50.351 --rc geninfo_unexecuted_blocks=1 00:21:50.351 00:21:50.351 ' 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:21:50.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:50.351 --rc genhtml_branch_coverage=1 00:21:50.351 --rc genhtml_function_coverage=1 00:21:50.351 --rc genhtml_legend=1 00:21:50.351 --rc geninfo_all_blocks=1 00:21:50.351 --rc geninfo_unexecuted_blocks=1 00:21:50.351 00:21:50.351 ' 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:21:50.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:50.351 --rc genhtml_branch_coverage=1 00:21:50.351 --rc genhtml_function_coverage=1 00:21:50.351 --rc genhtml_legend=1 00:21:50.351 --rc geninfo_all_blocks=1 00:21:50.351 --rc geninfo_unexecuted_blocks=1 00:21:50.351 00:21:50.351 ' 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # uname -s 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:21:50.351 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@15 -- # shopt -s extglob 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@5 -- # export PATH 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@51 -- # : 0 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:21:50.352 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@55 -- # have_pci_nics=0 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@12 -- # MALLOC_BDEV_SIZE=64 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@162 -- # run_test nvmf_shutdown_tc1 nvmf_shutdown_tc1 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:21:50.352 ************************************ 00:21:50.352 START TEST nvmf_shutdown_tc1 00:21:50.352 ************************************ 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc1 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@75 -- # starttarget 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@16 -- # nvmftestinit 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@474 -- # prepare_net_devs 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@309 -- # xtrace_disable 00:21:50.352 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # pci_devs=() 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # local -a pci_devs 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # pci_drivers=() 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # net_devs=() 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # local -ga net_devs 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # e810=() 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # local -ga e810 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # x722=() 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # local -ga x722 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # mlx=() 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # local -ga mlx 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:21:58.494 Found 0000:31:00.0 (0x8086 - 0x159b) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:21:58.494 Found 0000:31:00.1 (0x8086 - 0x159b) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:21:58.494 Found net devices under 0000:31:00.0: cvl_0_0 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:21:58.494 Found net devices under 0000:31:00.1: cvl_0_1 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # is_hw=yes 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:21:58.494 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:21:58.495 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:58.495 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.621 ms 00:21:58.495 00:21:58.495 --- 10.0.0.2 ping statistics --- 00:21:58.495 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:58.495 rtt min/avg/max/mdev = 0.621/0.621/0.621/0.000 ms 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:58.495 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:58.495 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.277 ms 00:21:58.495 00:21:58.495 --- 10.0.0.1 ping statistics --- 00:21:58.495 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:58.495 rtt min/avg/max/mdev = 0.277/0.277/0.277/0.000 ms 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@448 -- # return 0 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@724 -- # xtrace_disable 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@507 -- # nvmfpid=112206 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@508 -- # waitforlisten 112206 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@831 -- # '[' -z 112206 ']' 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:58.495 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:58.495 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:58.495 [2024-10-17 17:43:05.675369] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:21:58.495 [2024-10-17 17:43:05.675447] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:58.495 [2024-10-17 17:43:05.766416] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:58.495 [2024-10-17 17:43:05.818984] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:58.495 [2024-10-17 17:43:05.819040] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:58.495 [2024-10-17 17:43:05.819050] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:58.495 [2024-10-17 17:43:05.819057] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:58.495 [2024-10-17 17:43:05.819064] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:58.495 [2024-10-17 17:43:05.821204] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:21:58.495 [2024-10-17 17:43:05.821363] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:21:58.495 [2024-10-17 17:43:05.821528] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:58.495 [2024-10-17 17:43:05.821528] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # return 0 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@730 -- # xtrace_disable 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:58.756 [2024-10-17 17:43:06.521001] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@724 -- # xtrace_disable 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@36 -- # rpc_cmd 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:58.756 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:58.756 Malloc1 00:21:58.756 [2024-10-17 17:43:06.644306] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:58.756 Malloc2 00:21:59.017 Malloc3 00:21:59.017 Malloc4 00:21:59.017 Malloc5 00:21:59.017 Malloc6 00:21:59.017 Malloc7 00:21:59.017 Malloc8 00:21:59.280 Malloc9 00:21:59.280 Malloc10 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@730 -- # xtrace_disable 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@79 -- # perfpid=112585 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@80 -- # waitforlisten 112585 /var/tmp/bdevperf.sock 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@831 -- # '[' -z 112585 ']' 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json /dev/fd/63 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:59.280 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # config=() 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # local subsystem config 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.280 { 00:21:59.280 "params": { 00:21:59.280 "name": "Nvme$subsystem", 00:21:59.280 "trtype": "$TEST_TRANSPORT", 00:21:59.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.280 "adrfam": "ipv4", 00:21:59.280 "trsvcid": "$NVMF_PORT", 00:21:59.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.280 "hdgst": ${hdgst:-false}, 00:21:59.280 "ddgst": ${ddgst:-false} 00:21:59.280 }, 00:21:59.280 "method": "bdev_nvme_attach_controller" 00:21:59.280 } 00:21:59.280 EOF 00:21:59.280 )") 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.280 { 00:21:59.280 "params": { 00:21:59.280 "name": "Nvme$subsystem", 00:21:59.280 "trtype": "$TEST_TRANSPORT", 00:21:59.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.280 "adrfam": "ipv4", 00:21:59.280 "trsvcid": "$NVMF_PORT", 00:21:59.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.280 "hdgst": ${hdgst:-false}, 00:21:59.280 "ddgst": ${ddgst:-false} 00:21:59.280 }, 00:21:59.280 "method": "bdev_nvme_attach_controller" 00:21:59.280 } 00:21:59.280 EOF 00:21:59.280 )") 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.280 { 00:21:59.280 "params": { 00:21:59.280 "name": "Nvme$subsystem", 00:21:59.280 "trtype": "$TEST_TRANSPORT", 00:21:59.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.280 "adrfam": "ipv4", 00:21:59.280 "trsvcid": "$NVMF_PORT", 00:21:59.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.280 "hdgst": ${hdgst:-false}, 00:21:59.280 "ddgst": ${ddgst:-false} 00:21:59.280 }, 00:21:59.280 "method": "bdev_nvme_attach_controller" 00:21:59.280 } 00:21:59.280 EOF 00:21:59.280 )") 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.280 { 00:21:59.280 "params": { 00:21:59.280 "name": "Nvme$subsystem", 00:21:59.280 "trtype": "$TEST_TRANSPORT", 00:21:59.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.280 "adrfam": "ipv4", 00:21:59.280 "trsvcid": "$NVMF_PORT", 00:21:59.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.280 "hdgst": ${hdgst:-false}, 00:21:59.280 "ddgst": ${ddgst:-false} 00:21:59.280 }, 00:21:59.280 "method": "bdev_nvme_attach_controller" 00:21:59.280 } 00:21:59.280 EOF 00:21:59.280 )") 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.280 { 00:21:59.280 "params": { 00:21:59.280 "name": "Nvme$subsystem", 00:21:59.280 "trtype": "$TEST_TRANSPORT", 00:21:59.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.280 "adrfam": "ipv4", 00:21:59.280 "trsvcid": "$NVMF_PORT", 00:21:59.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.280 "hdgst": ${hdgst:-false}, 00:21:59.280 "ddgst": ${ddgst:-false} 00:21:59.280 }, 00:21:59.280 "method": "bdev_nvme_attach_controller" 00:21:59.280 } 00:21:59.280 EOF 00:21:59.280 )") 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.280 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.280 { 00:21:59.280 "params": { 00:21:59.281 "name": "Nvme$subsystem", 00:21:59.281 "trtype": "$TEST_TRANSPORT", 00:21:59.281 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "$NVMF_PORT", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.281 "hdgst": ${hdgst:-false}, 00:21:59.281 "ddgst": ${ddgst:-false} 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 } 00:21:59.281 EOF 00:21:59.281 )") 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.281 [2024-10-17 17:43:07.105217] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:21:59.281 [2024-10-17 17:43:07.105271] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.281 { 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme$subsystem", 00:21:59.281 "trtype": "$TEST_TRANSPORT", 00:21:59.281 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "$NVMF_PORT", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.281 "hdgst": ${hdgst:-false}, 00:21:59.281 "ddgst": ${ddgst:-false} 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 } 00:21:59.281 EOF 00:21:59.281 )") 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.281 { 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme$subsystem", 00:21:59.281 "trtype": "$TEST_TRANSPORT", 00:21:59.281 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "$NVMF_PORT", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.281 "hdgst": ${hdgst:-false}, 00:21:59.281 "ddgst": ${ddgst:-false} 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 } 00:21:59.281 EOF 00:21:59.281 )") 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.281 { 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme$subsystem", 00:21:59.281 "trtype": "$TEST_TRANSPORT", 00:21:59.281 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "$NVMF_PORT", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.281 "hdgst": ${hdgst:-false}, 00:21:59.281 "ddgst": ${ddgst:-false} 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 } 00:21:59.281 EOF 00:21:59.281 )") 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:21:59.281 { 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme$subsystem", 00:21:59.281 "trtype": "$TEST_TRANSPORT", 00:21:59.281 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "$NVMF_PORT", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:59.281 "hdgst": ${hdgst:-false}, 00:21:59.281 "ddgst": ${ddgst:-false} 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 } 00:21:59.281 EOF 00:21:59.281 )") 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # jq . 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@583 -- # IFS=, 00:21:59.281 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme1", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme2", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme3", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme4", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme5", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme6", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme7", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme8", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme9", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 },{ 00:21:59.281 "params": { 00:21:59.281 "name": "Nvme10", 00:21:59.281 "trtype": "tcp", 00:21:59.281 "traddr": "10.0.0.2", 00:21:59.281 "adrfam": "ipv4", 00:21:59.281 "trsvcid": "4420", 00:21:59.281 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:21:59.281 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:21:59.281 "hdgst": false, 00:21:59.281 "ddgst": false 00:21:59.281 }, 00:21:59.281 "method": "bdev_nvme_attach_controller" 00:21:59.281 }' 00:21:59.281 [2024-10-17 17:43:07.183496] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:59.542 [2024-10-17 17:43:07.220401] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # return 0 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@81 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@84 -- # kill -9 112585 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@85 -- # rm -f /var/run/spdk_bdev1 00:22:00.926 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 74: 112585 Killed $rootdir/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json <(gen_nvmf_target_json "${num_subsystems[@]}") 00:22:00.926 17:43:08 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@88 -- # sleep 1 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@89 -- # kill -0 112206 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # config=() 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # local subsystem config 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.869 "ddgst": ${ddgst:-false} 00:22:01.869 }, 00:22:01.869 "method": "bdev_nvme_attach_controller" 00:22:01.869 } 00:22:01.869 EOF 00:22:01.869 )") 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.869 "ddgst": ${ddgst:-false} 00:22:01.869 }, 00:22:01.869 "method": "bdev_nvme_attach_controller" 00:22:01.869 } 00:22:01.869 EOF 00:22:01.869 )") 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.869 "ddgst": ${ddgst:-false} 00:22:01.869 }, 00:22:01.869 "method": "bdev_nvme_attach_controller" 00:22:01.869 } 00:22:01.869 EOF 00:22:01.869 )") 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.869 "ddgst": ${ddgst:-false} 00:22:01.869 }, 00:22:01.869 "method": "bdev_nvme_attach_controller" 00:22:01.869 } 00:22:01.869 EOF 00:22:01.869 )") 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.869 "ddgst": ${ddgst:-false} 00:22:01.869 }, 00:22:01.869 "method": "bdev_nvme_attach_controller" 00:22:01.869 } 00:22:01.869 EOF 00:22:01.869 )") 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.869 "ddgst": ${ddgst:-false} 00:22:01.869 }, 00:22:01.869 "method": "bdev_nvme_attach_controller" 00:22:01.869 } 00:22:01.869 EOF 00:22:01.869 )") 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.869 [2024-10-17 17:43:09.630450] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:01.869 [2024-10-17 17:43:09.630499] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid113144 ] 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.869 "ddgst": ${ddgst:-false} 00:22:01.869 }, 00:22:01.869 "method": "bdev_nvme_attach_controller" 00:22:01.869 } 00:22:01.869 EOF 00:22:01.869 )") 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.869 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.869 { 00:22:01.869 "params": { 00:22:01.869 "name": "Nvme$subsystem", 00:22:01.869 "trtype": "$TEST_TRANSPORT", 00:22:01.869 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.869 "adrfam": "ipv4", 00:22:01.869 "trsvcid": "$NVMF_PORT", 00:22:01.869 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.869 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.869 "hdgst": ${hdgst:-false}, 00:22:01.870 "ddgst": ${ddgst:-false} 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 } 00:22:01.870 EOF 00:22:01.870 )") 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.870 { 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme$subsystem", 00:22:01.870 "trtype": "$TEST_TRANSPORT", 00:22:01.870 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "$NVMF_PORT", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.870 "hdgst": ${hdgst:-false}, 00:22:01.870 "ddgst": ${ddgst:-false} 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 } 00:22:01.870 EOF 00:22:01.870 )") 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:01.870 { 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme$subsystem", 00:22:01.870 "trtype": "$TEST_TRANSPORT", 00:22:01.870 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "$NVMF_PORT", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:01.870 "hdgst": ${hdgst:-false}, 00:22:01.870 "ddgst": ${ddgst:-false} 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 } 00:22:01.870 EOF 00:22:01.870 )") 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # jq . 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@583 -- # IFS=, 00:22:01.870 17:43:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme1", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme2", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme3", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme4", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme5", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme6", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme7", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme8", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme9", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 },{ 00:22:01.870 "params": { 00:22:01.870 "name": "Nvme10", 00:22:01.870 "trtype": "tcp", 00:22:01.870 "traddr": "10.0.0.2", 00:22:01.870 "adrfam": "ipv4", 00:22:01.870 "trsvcid": "4420", 00:22:01.870 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:22:01.870 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:22:01.870 "hdgst": false, 00:22:01.870 "ddgst": false 00:22:01.870 }, 00:22:01.870 "method": "bdev_nvme_attach_controller" 00:22:01.870 }' 00:22:01.870 [2024-10-17 17:43:09.710345] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:01.870 [2024-10-17 17:43:09.746355] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:03.256 Running I/O for 1 seconds... 00:22:04.458 1847.00 IOPS, 115.44 MiB/s 00:22:04.458 Latency(us) 00:22:04.458 [2024-10-17T15:43:12.377Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:04.458 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme1n1 : 1.03 186.73 11.67 0.00 0.00 337825.56 19223.89 281367.89 00:22:04.458 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme2n1 : 1.15 222.22 13.89 0.00 0.00 278110.72 16820.91 248162.99 00:22:04.458 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme3n1 : 1.11 230.58 14.41 0.00 0.00 259146.67 16165.55 248162.99 00:22:04.458 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme4n1 : 1.19 270.00 16.87 0.00 0.00 221345.37 13489.49 242920.11 00:22:04.458 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme5n1 : 1.16 221.43 13.84 0.00 0.00 265934.51 17694.72 248162.99 00:22:04.458 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme6n1 : 1.19 267.95 16.75 0.00 0.00 216500.39 18240.85 241172.48 00:22:04.458 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme7n1 : 1.18 274.44 17.15 0.00 0.00 206500.47 2990.08 249910.61 00:22:04.458 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme8n1 : 1.14 224.00 14.00 0.00 0.00 248469.76 20316.16 246415.36 00:22:04.458 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme9n1 : 1.19 272.03 17.00 0.00 0.00 201813.67 1037.65 227191.47 00:22:04.458 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:04.458 Verification LBA range: start 0x0 length 0x400 00:22:04.458 Nvme10n1 : 1.20 266.07 16.63 0.00 0.00 202721.54 8246.61 265639.25 00:22:04.458 [2024-10-17T15:43:12.377Z] =================================================================================================================== 00:22:04.458 [2024-10-17T15:43:12.377Z] Total : 2435.43 152.21 0.00 0.00 237734.85 1037.65 281367.89 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@95 -- # stoptarget 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@46 -- # nvmftestfini 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@121 -- # sync 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@124 -- # set +e 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:04.458 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:04.458 rmmod nvme_tcp 00:22:04.458 rmmod nvme_fabrics 00:22:04.458 rmmod nvme_keyring 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@128 -- # set -e 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@129 -- # return 0 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@515 -- # '[' -n 112206 ']' 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@516 -- # killprocess 112206 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@950 -- # '[' -z 112206 ']' 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@954 -- # kill -0 112206 00:22:04.718 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@955 -- # uname 00:22:04.719 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:04.719 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 112206 00:22:04.719 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:22:04.719 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:22:04.719 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 112206' 00:22:04.719 killing process with pid 112206 00:22:04.719 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@969 -- # kill 112206 00:22:04.719 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@974 -- # wait 112206 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@297 -- # iptr 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@789 -- # iptables-save 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@789 -- # iptables-restore 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:04.980 17:43:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:06.895 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:06.895 00:22:06.895 real 0m16.541s 00:22:06.895 user 0m33.422s 00:22:06.895 sys 0m6.641s 00:22:06.895 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:06.895 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:22:06.895 ************************************ 00:22:06.895 END TEST nvmf_shutdown_tc1 00:22:06.895 ************************************ 00:22:06.895 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@163 -- # run_test nvmf_shutdown_tc2 nvmf_shutdown_tc2 00:22:06.895 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:22:06.895 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:06.895 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:22:07.157 ************************************ 00:22:07.157 START TEST nvmf_shutdown_tc2 00:22:07.157 ************************************ 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc2 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@100 -- # starttarget 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@16 -- # nvmftestinit 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@474 -- # prepare_net_devs 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@309 -- # xtrace_disable 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # pci_devs=() 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # net_devs=() 00:22:07.157 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # e810=() 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # local -ga e810 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # x722=() 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # local -ga x722 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # mlx=() 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # local -ga mlx 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:07.158 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:07.158 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:07.158 Found net devices under 0000:31:00.0: cvl_0_0 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:07.158 Found net devices under 0000:31:00.1: cvl_0_1 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # is_hw=yes 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:07.158 17:43:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:07.158 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:07.158 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:07.158 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:07.158 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:07.419 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:07.419 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.618 ms 00:22:07.419 00:22:07.419 --- 10.0.0.2 ping statistics --- 00:22:07.419 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:07.419 rtt min/avg/max/mdev = 0.618/0.618/0.618/0.000 ms 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:07.419 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:07.419 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.289 ms 00:22:07.419 00:22:07.419 --- 10.0.0.1 ping statistics --- 00:22:07.419 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:07.419 rtt min/avg/max/mdev = 0.289/0.289/0.289/0.000 ms 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@448 -- # return 0 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:22:07.419 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@507 -- # nvmfpid=114388 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@508 -- # waitforlisten 114388 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@831 -- # '[' -z 114388 ']' 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:07.420 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:07.420 17:43:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:07.420 [2024-10-17 17:43:15.318862] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:07.420 [2024-10-17 17:43:15.318912] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:07.680 [2024-10-17 17:43:15.405054] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:07.680 [2024-10-17 17:43:15.436872] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:07.680 [2024-10-17 17:43:15.436902] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:07.680 [2024-10-17 17:43:15.436908] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:07.681 [2024-10-17 17:43:15.436916] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:07.681 [2024-10-17 17:43:15.436921] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:07.681 [2024-10-17 17:43:15.438478] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:07.681 [2024-10-17 17:43:15.438634] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:07.681 [2024-10-17 17:43:15.438784] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:22:07.681 [2024-10-17 17:43:15.438935] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # return 0 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:08.252 [2024-10-17 17:43:16.160239] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:08.252 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@36 -- # rpc_cmd 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.513 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:08.513 Malloc1 00:22:08.513 [2024-10-17 17:43:16.270530] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:08.513 Malloc2 00:22:08.513 Malloc3 00:22:08.513 Malloc4 00:22:08.513 Malloc5 00:22:08.774 Malloc6 00:22:08.774 Malloc7 00:22:08.774 Malloc8 00:22:08.774 Malloc9 00:22:08.774 Malloc10 00:22:08.774 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.774 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:22:08.774 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:08.774 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:08.774 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@104 -- # perfpid=114645 00:22:08.774 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@105 -- # waitforlisten 114645 /var/tmp/bdevperf.sock 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@831 -- # '[' -z 114645 ']' 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:22:08.775 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@558 -- # config=() 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@558 -- # local subsystem config 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:08.775 { 00:22:08.775 "params": { 00:22:08.775 "name": "Nvme$subsystem", 00:22:08.775 "trtype": "$TEST_TRANSPORT", 00:22:08.775 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:08.775 "adrfam": "ipv4", 00:22:08.775 "trsvcid": "$NVMF_PORT", 00:22:08.775 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:08.775 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:08.775 "hdgst": ${hdgst:-false}, 00:22:08.775 "ddgst": ${ddgst:-false} 00:22:08.775 }, 00:22:08.775 "method": "bdev_nvme_attach_controller" 00:22:08.775 } 00:22:08.775 EOF 00:22:08.775 )") 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:08.775 { 00:22:08.775 "params": { 00:22:08.775 "name": "Nvme$subsystem", 00:22:08.775 "trtype": "$TEST_TRANSPORT", 00:22:08.775 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:08.775 "adrfam": "ipv4", 00:22:08.775 "trsvcid": "$NVMF_PORT", 00:22:08.775 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:08.775 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:08.775 "hdgst": ${hdgst:-false}, 00:22:08.775 "ddgst": ${ddgst:-false} 00:22:08.775 }, 00:22:08.775 "method": "bdev_nvme_attach_controller" 00:22:08.775 } 00:22:08.775 EOF 00:22:08.775 )") 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:08.775 { 00:22:08.775 "params": { 00:22:08.775 "name": "Nvme$subsystem", 00:22:08.775 "trtype": "$TEST_TRANSPORT", 00:22:08.775 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:08.775 "adrfam": "ipv4", 00:22:08.775 "trsvcid": "$NVMF_PORT", 00:22:08.775 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:08.775 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:08.775 "hdgst": ${hdgst:-false}, 00:22:08.775 "ddgst": ${ddgst:-false} 00:22:08.775 }, 00:22:08.775 "method": "bdev_nvme_attach_controller" 00:22:08.775 } 00:22:08.775 EOF 00:22:08.775 )") 00:22:08.775 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:09.036 { 00:22:09.036 "params": { 00:22:09.036 "name": "Nvme$subsystem", 00:22:09.036 "trtype": "$TEST_TRANSPORT", 00:22:09.036 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:09.036 "adrfam": "ipv4", 00:22:09.036 "trsvcid": "$NVMF_PORT", 00:22:09.036 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:09.036 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:09.036 "hdgst": ${hdgst:-false}, 00:22:09.036 "ddgst": ${ddgst:-false} 00:22:09.036 }, 00:22:09.036 "method": "bdev_nvme_attach_controller" 00:22:09.036 } 00:22:09.036 EOF 00:22:09.036 )") 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:09.036 { 00:22:09.036 "params": { 00:22:09.036 "name": "Nvme$subsystem", 00:22:09.036 "trtype": "$TEST_TRANSPORT", 00:22:09.036 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:09.036 "adrfam": "ipv4", 00:22:09.036 "trsvcid": "$NVMF_PORT", 00:22:09.036 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:09.036 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:09.036 "hdgst": ${hdgst:-false}, 00:22:09.036 "ddgst": ${ddgst:-false} 00:22:09.036 }, 00:22:09.036 "method": "bdev_nvme_attach_controller" 00:22:09.036 } 00:22:09.036 EOF 00:22:09.036 )") 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:09.036 { 00:22:09.036 "params": { 00:22:09.036 "name": "Nvme$subsystem", 00:22:09.036 "trtype": "$TEST_TRANSPORT", 00:22:09.036 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:09.036 "adrfam": "ipv4", 00:22:09.036 "trsvcid": "$NVMF_PORT", 00:22:09.036 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:09.036 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:09.036 "hdgst": ${hdgst:-false}, 00:22:09.036 "ddgst": ${ddgst:-false} 00:22:09.036 }, 00:22:09.036 "method": "bdev_nvme_attach_controller" 00:22:09.036 } 00:22:09.036 EOF 00:22:09.036 )") 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.036 [2024-10-17 17:43:16.718501] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:09.036 [2024-10-17 17:43:16.718556] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid114645 ] 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:09.036 { 00:22:09.036 "params": { 00:22:09.036 "name": "Nvme$subsystem", 00:22:09.036 "trtype": "$TEST_TRANSPORT", 00:22:09.036 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:09.036 "adrfam": "ipv4", 00:22:09.036 "trsvcid": "$NVMF_PORT", 00:22:09.036 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:09.036 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:09.036 "hdgst": ${hdgst:-false}, 00:22:09.036 "ddgst": ${ddgst:-false} 00:22:09.036 }, 00:22:09.036 "method": "bdev_nvme_attach_controller" 00:22:09.036 } 00:22:09.036 EOF 00:22:09.036 )") 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:09.036 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:09.037 { 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme$subsystem", 00:22:09.037 "trtype": "$TEST_TRANSPORT", 00:22:09.037 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "$NVMF_PORT", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:09.037 "hdgst": ${hdgst:-false}, 00:22:09.037 "ddgst": ${ddgst:-false} 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 } 00:22:09.037 EOF 00:22:09.037 )") 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:09.037 { 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme$subsystem", 00:22:09.037 "trtype": "$TEST_TRANSPORT", 00:22:09.037 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "$NVMF_PORT", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:09.037 "hdgst": ${hdgst:-false}, 00:22:09.037 "ddgst": ${ddgst:-false} 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 } 00:22:09.037 EOF 00:22:09.037 )") 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:09.037 { 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme$subsystem", 00:22:09.037 "trtype": "$TEST_TRANSPORT", 00:22:09.037 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "$NVMF_PORT", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:09.037 "hdgst": ${hdgst:-false}, 00:22:09.037 "ddgst": ${ddgst:-false} 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 } 00:22:09.037 EOF 00:22:09.037 )") 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # jq . 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@583 -- # IFS=, 00:22:09.037 17:43:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme1", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme2", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme3", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme4", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme5", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme6", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme7", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme8", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme9", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 },{ 00:22:09.037 "params": { 00:22:09.037 "name": "Nvme10", 00:22:09.037 "trtype": "tcp", 00:22:09.037 "traddr": "10.0.0.2", 00:22:09.037 "adrfam": "ipv4", 00:22:09.037 "trsvcid": "4420", 00:22:09.037 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:22:09.037 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:22:09.037 "hdgst": false, 00:22:09.037 "ddgst": false 00:22:09.037 }, 00:22:09.037 "method": "bdev_nvme_attach_controller" 00:22:09.037 }' 00:22:09.037 [2024-10-17 17:43:16.796427] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:09.037 [2024-10-17 17:43:16.833161] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:10.419 Running I/O for 10 seconds... 00:22:10.419 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:10.419 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # return 0 00:22:10.419 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@106 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:22:10.419 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.419 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@108 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@58 -- # local ret=1 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@59 -- # local i 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=3 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 3 -ge 100 ']' 00:22:10.679 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@68 -- # sleep 0.25 00:22:10.939 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i-- )) 00:22:10.939 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:22:10.939 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:22:10.939 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:22:10.939 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.940 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:10.940 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.940 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=67 00:22:10.940 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 67 -ge 100 ']' 00:22:10.940 17:43:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@68 -- # sleep 0.25 00:22:11.200 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i-- )) 00:22:11.200 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:22:11.200 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:22:11.200 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:22:11.200 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.200 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:11.200 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=131 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 131 -ge 100 ']' 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@65 -- # ret=0 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@66 -- # break 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@70 -- # return 0 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@111 -- # killprocess 114645 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@950 -- # '[' -z 114645 ']' 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # kill -0 114645 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # uname 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 114645 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 114645' 00:22:11.461 killing process with pid 114645 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@969 -- # kill 114645 00:22:11.461 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@974 -- # wait 114645 00:22:11.461 Received shutdown signal, test time was about 0.984305 seconds 00:22:11.461 00:22:11.461 Latency(us) 00:22:11.461 [2024-10-17T15:43:19.380Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:11.461 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme1n1 : 0.98 262.01 16.38 0.00 0.00 241472.85 19114.67 253405.87 00:22:11.461 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme2n1 : 0.97 268.65 16.79 0.00 0.00 230421.99 672.43 246415.36 00:22:11.461 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme3n1 : 0.98 261.13 16.32 0.00 0.00 232579.84 19005.44 248162.99 00:22:11.461 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme4n1 : 0.98 260.32 16.27 0.00 0.00 228342.40 14527.15 251658.24 00:22:11.461 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme5n1 : 0.97 262.91 16.43 0.00 0.00 221206.83 21408.43 263891.63 00:22:11.461 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme6n1 : 0.95 201.25 12.58 0.00 0.00 281860.84 19660.80 255153.49 00:22:11.461 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme7n1 : 0.97 268.67 16.79 0.00 0.00 205707.53 5406.72 223696.21 00:22:11.461 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme8n1 : 0.95 202.92 12.68 0.00 0.00 266379.66 15291.73 249910.61 00:22:11.461 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme9n1 : 0.96 200.26 12.52 0.00 0.00 264034.70 17913.17 251658.24 00:22:11.461 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:11.461 Verification LBA range: start 0x0 length 0x400 00:22:11.461 Nvme10n1 : 0.97 198.88 12.43 0.00 0.00 260072.11 19879.25 272629.76 00:22:11.461 [2024-10-17T15:43:19.380Z] =================================================================================================================== 00:22:11.461 [2024-10-17T15:43:19.380Z] Total : 2387.00 149.19 0.00 0.00 240361.81 672.43 272629.76 00:22:11.722 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@114 -- # sleep 1 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@115 -- # kill -0 114388 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@117 -- # stoptarget 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@46 -- # nvmftestfini 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@121 -- # sync 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@124 -- # set +e 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:12.665 rmmod nvme_tcp 00:22:12.665 rmmod nvme_fabrics 00:22:12.665 rmmod nvme_keyring 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@128 -- # set -e 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@129 -- # return 0 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@515 -- # '[' -n 114388 ']' 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@516 -- # killprocess 114388 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@950 -- # '[' -z 114388 ']' 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # kill -0 114388 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # uname 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 114388 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 114388' 00:22:12.665 killing process with pid 114388 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@969 -- # kill 114388 00:22:12.665 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@974 -- # wait 114388 00:22:12.925 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:12.925 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:12.925 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:12.925 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@297 -- # iptr 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@789 -- # iptables-save 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@789 -- # iptables-restore 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:12.926 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:15.473 00:22:15.473 real 0m8.003s 00:22:15.473 user 0m24.213s 00:22:15.473 sys 0m1.318s 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:22:15.473 ************************************ 00:22:15.473 END TEST nvmf_shutdown_tc2 00:22:15.473 ************************************ 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@164 -- # run_test nvmf_shutdown_tc3 nvmf_shutdown_tc3 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:22:15.473 ************************************ 00:22:15.473 START TEST nvmf_shutdown_tc3 00:22:15.473 ************************************ 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc3 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@122 -- # starttarget 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@16 -- # nvmftestinit 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@474 -- # prepare_net_devs 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@309 -- # xtrace_disable 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # pci_devs=() 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # net_devs=() 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # e810=() 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # local -ga e810 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # x722=() 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # local -ga x722 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # mlx=() 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # local -ga mlx 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:15.473 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:15.473 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:15.473 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:15.474 Found net devices under 0000:31:00.0: cvl_0_0 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:15.474 Found net devices under 0000:31:00.1: cvl_0_1 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # is_hw=yes 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:15.474 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:15.474 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:15.474 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.610 ms 00:22:15.474 00:22:15.474 --- 10.0.0.2 ping statistics --- 00:22:15.474 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:15.474 rtt min/avg/max/mdev = 0.610/0.610/0.610/0.000 ms 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:15.474 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:15.474 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.194 ms 00:22:15.474 00:22:15.474 --- 10.0.0.1 ping statistics --- 00:22:15.474 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:15.474 rtt min/avg/max/mdev = 0.194/0.194/0.194/0.000 ms 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@448 -- # return 0 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@507 -- # nvmfpid=115930 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@508 -- # waitforlisten 115930 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@831 -- # '[' -z 115930 ']' 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:15.474 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:15.474 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:15.736 [2024-10-17 17:43:23.402101] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:15.736 [2024-10-17 17:43:23.402162] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:15.736 [2024-10-17 17:43:23.489671] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:15.736 [2024-10-17 17:43:23.522491] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:15.736 [2024-10-17 17:43:23.522521] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:15.736 [2024-10-17 17:43:23.522526] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:15.736 [2024-10-17 17:43:23.522535] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:15.736 [2024-10-17 17:43:23.522539] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:15.736 [2024-10-17 17:43:23.523879] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:15.736 [2024-10-17 17:43:23.524083] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:15.736 [2024-10-17 17:43:23.524237] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:15.736 [2024-10-17 17:43:23.524238] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:22:16.307 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:16.307 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # return 0 00:22:16.307 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:22:16.307 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:16.307 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:16.568 [2024-10-17 17:43:24.249509] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@36 -- # rpc_cmd 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:16.568 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:16.568 Malloc1 00:22:16.568 [2024-10-17 17:43:24.359620] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:16.568 Malloc2 00:22:16.568 Malloc3 00:22:16.568 Malloc4 00:22:16.828 Malloc5 00:22:16.828 Malloc6 00:22:16.828 Malloc7 00:22:16.828 Malloc8 00:22:16.828 Malloc9 00:22:16.828 Malloc10 00:22:16.828 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:16.828 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:22:16.828 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:16.828 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@126 -- # perfpid=116296 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@127 -- # waitforlisten 116296 /var/tmp/bdevperf.sock 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@831 -- # '[' -z 116296 ']' 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:22:17.089 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@558 -- # config=() 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@558 -- # local subsystem config 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.089 { 00:22:17.089 "params": { 00:22:17.089 "name": "Nvme$subsystem", 00:22:17.089 "trtype": "$TEST_TRANSPORT", 00:22:17.089 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.089 "adrfam": "ipv4", 00:22:17.089 "trsvcid": "$NVMF_PORT", 00:22:17.089 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.089 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.089 "hdgst": ${hdgst:-false}, 00:22:17.089 "ddgst": ${ddgst:-false} 00:22:17.089 }, 00:22:17.089 "method": "bdev_nvme_attach_controller" 00:22:17.089 } 00:22:17.089 EOF 00:22:17.089 )") 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.089 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.089 { 00:22:17.089 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 [2024-10-17 17:43:24.816720] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:17.090 [2024-10-17 17:43:24.816789] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid116296 ] 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:17.090 { 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme$subsystem", 00:22:17.090 "trtype": "$TEST_TRANSPORT", 00:22:17.090 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "$NVMF_PORT", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:17.090 "hdgst": ${hdgst:-false}, 00:22:17.090 "ddgst": ${ddgst:-false} 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 } 00:22:17.090 EOF 00:22:17.090 )") 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # jq . 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@583 -- # IFS=, 00:22:17.090 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme1", 00:22:17.090 "trtype": "tcp", 00:22:17.090 "traddr": "10.0.0.2", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "4420", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:22:17.090 "hdgst": false, 00:22:17.090 "ddgst": false 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 },{ 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme2", 00:22:17.090 "trtype": "tcp", 00:22:17.090 "traddr": "10.0.0.2", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "4420", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:22:17.090 "hdgst": false, 00:22:17.090 "ddgst": false 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 },{ 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme3", 00:22:17.090 "trtype": "tcp", 00:22:17.090 "traddr": "10.0.0.2", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "4420", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:22:17.090 "hdgst": false, 00:22:17.090 "ddgst": false 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.090 },{ 00:22:17.090 "params": { 00:22:17.090 "name": "Nvme4", 00:22:17.090 "trtype": "tcp", 00:22:17.090 "traddr": "10.0.0.2", 00:22:17.090 "adrfam": "ipv4", 00:22:17.090 "trsvcid": "4420", 00:22:17.090 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:22:17.090 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:22:17.090 "hdgst": false, 00:22:17.090 "ddgst": false 00:22:17.090 }, 00:22:17.090 "method": "bdev_nvme_attach_controller" 00:22:17.091 },{ 00:22:17.091 "params": { 00:22:17.091 "name": "Nvme5", 00:22:17.091 "trtype": "tcp", 00:22:17.091 "traddr": "10.0.0.2", 00:22:17.091 "adrfam": "ipv4", 00:22:17.091 "trsvcid": "4420", 00:22:17.091 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:22:17.091 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:22:17.091 "hdgst": false, 00:22:17.091 "ddgst": false 00:22:17.091 }, 00:22:17.091 "method": "bdev_nvme_attach_controller" 00:22:17.091 },{ 00:22:17.091 "params": { 00:22:17.091 "name": "Nvme6", 00:22:17.091 "trtype": "tcp", 00:22:17.091 "traddr": "10.0.0.2", 00:22:17.091 "adrfam": "ipv4", 00:22:17.091 "trsvcid": "4420", 00:22:17.091 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:22:17.091 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:22:17.091 "hdgst": false, 00:22:17.091 "ddgst": false 00:22:17.091 }, 00:22:17.091 "method": "bdev_nvme_attach_controller" 00:22:17.091 },{ 00:22:17.091 "params": { 00:22:17.091 "name": "Nvme7", 00:22:17.091 "trtype": "tcp", 00:22:17.091 "traddr": "10.0.0.2", 00:22:17.091 "adrfam": "ipv4", 00:22:17.091 "trsvcid": "4420", 00:22:17.091 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:22:17.091 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:22:17.091 "hdgst": false, 00:22:17.091 "ddgst": false 00:22:17.091 }, 00:22:17.091 "method": "bdev_nvme_attach_controller" 00:22:17.091 },{ 00:22:17.091 "params": { 00:22:17.091 "name": "Nvme8", 00:22:17.091 "trtype": "tcp", 00:22:17.091 "traddr": "10.0.0.2", 00:22:17.091 "adrfam": "ipv4", 00:22:17.091 "trsvcid": "4420", 00:22:17.091 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:22:17.091 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:22:17.091 "hdgst": false, 00:22:17.091 "ddgst": false 00:22:17.091 }, 00:22:17.091 "method": "bdev_nvme_attach_controller" 00:22:17.091 },{ 00:22:17.091 "params": { 00:22:17.091 "name": "Nvme9", 00:22:17.091 "trtype": "tcp", 00:22:17.091 "traddr": "10.0.0.2", 00:22:17.091 "adrfam": "ipv4", 00:22:17.091 "trsvcid": "4420", 00:22:17.091 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:22:17.091 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:22:17.091 "hdgst": false, 00:22:17.091 "ddgst": false 00:22:17.091 }, 00:22:17.091 "method": "bdev_nvme_attach_controller" 00:22:17.091 },{ 00:22:17.091 "params": { 00:22:17.091 "name": "Nvme10", 00:22:17.091 "trtype": "tcp", 00:22:17.091 "traddr": "10.0.0.2", 00:22:17.091 "adrfam": "ipv4", 00:22:17.091 "trsvcid": "4420", 00:22:17.091 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:22:17.091 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:22:17.091 "hdgst": false, 00:22:17.091 "ddgst": false 00:22:17.091 }, 00:22:17.091 "method": "bdev_nvme_attach_controller" 00:22:17.091 }' 00:22:17.091 [2024-10-17 17:43:24.896840] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:17.091 [2024-10-17 17:43:24.933278] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:18.475 Running I/O for 10 seconds... 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # return 0 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@128 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@131 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@133 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@58 -- # local ret=1 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # local i 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:18.475 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.735 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=3 00:22:18.735 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 3 -ge 100 ']' 00:22:18.735 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:22:18.995 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:22:18.995 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:22:18.995 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:22:18.995 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:22:18.995 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.996 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:18.996 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.996 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=67 00:22:18.996 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 67 -ge 100 ']' 00:22:18.996 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:22:19.273 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:22:19.273 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:22:19.273 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:22:19.273 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:22:19.273 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.273 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:19.273 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=131 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 131 -ge 100 ']' 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@65 -- # ret=0 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@66 -- # break 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@70 -- # return 0 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@136 -- # killprocess 115930 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@950 -- # '[' -z 115930 ']' 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # kill -0 115930 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@955 -- # uname 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 115930 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 115930' 00:22:19.273 killing process with pid 115930 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@969 -- # kill 115930 00:22:19.273 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@974 -- # wait 115930 00:22:19.273 [2024-10-17 17:43:27.083169] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083218] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083224] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083230] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083235] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083240] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083244] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083249] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083254] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083259] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083263] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083268] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083273] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083277] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083282] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083286] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083291] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083296] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083306] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083311] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083317] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083321] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.273 [2024-10-17 17:43:27.083326] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083330] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083335] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083340] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083344] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083349] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083353] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083358] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083362] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083367] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083371] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083376] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083381] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083385] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083390] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083394] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083399] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083403] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083408] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083413] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083418] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083422] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083426] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083432] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083438] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083443] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083447] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083452] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083457] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083462] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083466] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083471] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083475] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083480] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083485] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083489] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083494] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083498] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083502] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083507] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.083511] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1291f80 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084481] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084508] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084514] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084520] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084525] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084530] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084536] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084540] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084545] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084553] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084557] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084562] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084567] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084572] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084576] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084582] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084587] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084592] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084597] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084602] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084607] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084612] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084616] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084622] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084626] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084631] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084637] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084642] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084647] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084652] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084657] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084662] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084666] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084671] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084675] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084680] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084686] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084695] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084700] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084705] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084709] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084714] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084719] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084724] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084729] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084734] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084739] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084744] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084748] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.274 [2024-10-17 17:43:27.084753] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084758] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084763] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084767] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084772] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084776] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084782] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084786] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084791] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084796] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084800] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084808] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084813] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.084818] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293d60 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.085348] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.275 [2024-10-17 17:43:27.085382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.275 [2024-10-17 17:43:27.085393] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.275 [2024-10-17 17:43:27.085402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.275 [2024-10-17 17:43:27.085410] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.275 [2024-10-17 17:43:27.085418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.275 [2024-10-17 17:43:27.085426] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.275 [2024-10-17 17:43:27.085434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.275 [2024-10-17 17:43:27.085441] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd26480 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086095] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086111] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086116] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086121] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086126] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086130] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086136] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086141] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086146] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086151] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086156] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086161] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086166] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086170] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086175] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086179] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086184] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086189] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086197] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086202] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086207] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086212] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086217] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086221] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086226] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086230] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086235] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086242] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086247] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086252] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086257] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086261] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086265] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086270] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086274] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086279] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086283] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086288] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086293] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086298] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086303] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086307] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086312] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086317] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086321] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086327] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086332] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086336] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086341] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086346] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086351] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086355] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086360] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086365] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086370] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086376] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086381] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086385] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086390] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086395] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086400] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086405] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.086410] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292450 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.087482] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.087506] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.087513] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.275 [2024-10-17 17:43:27.087518] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087523] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087528] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087533] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087538] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087543] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087553] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087558] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087563] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087569] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087573] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087578] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087583] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087588] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087593] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087597] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087603] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087609] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087613] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087619] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087623] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087628] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087633] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087637] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087642] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087647] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087651] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087656] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087661] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087667] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087671] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087676] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087681] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087687] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087696] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087701] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087706] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087711] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087716] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087721] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087726] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087731] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087735] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087741] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087745] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087750] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087754] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087759] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087764] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087769] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087774] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087779] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087784] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.087789] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292920 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.088662] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292e10 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.088685] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1292e10 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089230] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089245] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089251] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089256] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089267] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089272] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089278] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089283] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089288] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089293] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089298] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089302] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089307] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089312] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089317] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089322] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089327] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089332] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089337] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089343] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089348] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089353] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089357] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089362] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089367] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089372] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089377] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089383] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089388] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089393] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089398] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089403] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089409] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.276 [2024-10-17 17:43:27.089414] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089419] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089425] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089430] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089434] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089440] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089444] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089449] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089454] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089459] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089463] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089468] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089474] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089478] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089483] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089488] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089493] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089498] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089502] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089507] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089511] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089516] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089522] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089527] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089532] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089537] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089542] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089547] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089552] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.089557] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12932e0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090174] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023620 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090623] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090639] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090644] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090650] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090655] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090664] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090669] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090674] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090679] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090684] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090689] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090698] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090703] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090707] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090712] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090717] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090721] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090726] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090731] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090736] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090740] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090750] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090755] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090759] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090765] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090769] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090774] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090779] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090783] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090788] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090792] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090797] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090802] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090806] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090811] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090815] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090820] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090824] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090829] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090834] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090838] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090843] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.277 [2024-10-17 17:43:27.090848] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090852] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090857] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090861] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090866] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090871] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090879] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090884] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090888] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090893] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090898] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090902] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.090907] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1023af0 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091556] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091571] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091576] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091581] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091585] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091591] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091596] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091601] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091606] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091611] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091616] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091620] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091625] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091630] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091634] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091639] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091644] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091649] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091654] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091668] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091672] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091677] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091682] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091686] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091694] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091699] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091704] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091709] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091714] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091719] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:26240 len:1[2024-10-17 17:43:27.091725] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091736] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091740] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091745] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091750] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091755] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:26368 len:1[2024-10-17 17:43:27.091760] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091768] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091773] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091778] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.278 [2024-10-17 17:43:27.091781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.091986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.091996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.092003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.092013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.092023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.278 [2024-10-17 17:43:27.092033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.278 [2024-10-17 17:43:27.092041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.279 [2024-10-17 17:43:27.092769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.279 [2024-10-17 17:43:27.092776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.092786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.092794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.092804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.092812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.092822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.092829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.092838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.092845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.092856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.092863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.092916] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x11e2380 was disconnected and freed. reset controller. 00:22:19.280 [2024-10-17 17:43:27.093405] nvme_tcp.c:1252:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:22:19.280 [2024-10-17 17:43:27.094823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.094863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.094884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.094904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.094925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.094946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.094967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.094987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.094996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095413] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.280 [2024-10-17 17:43:27.095482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.280 [2024-10-17 17:43:27.095491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.095961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.281 [2024-10-17 17:43:27.095970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096018] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x113da60 was disconnected and freed. reset controller. 00:22:19.281 [2024-10-17 17:43:27.096152] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4] resetting controller 00:22:19.281 [2024-10-17 17:43:27.096199] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd23800 (9): Bad file descriptor 00:22:19.281 [2024-10-17 17:43:27.096223] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096242] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096258] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096274] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096289] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1165950 is same with the state(6) to be set 00:22:19.281 [2024-10-17 17:43:27.096315] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096341] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096360] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096376] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.281 [2024-10-17 17:43:27.096384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.281 [2024-10-17 17:43:27.096391] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1178a30 is same with the state(6) to be set 00:22:19.281 [2024-10-17 17:43:27.100907] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100928] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100935] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100945] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100950] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100955] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100960] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100965] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100971] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100976] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100981] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100986] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100991] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.100995] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101000] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101005] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101010] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101015] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101020] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101024] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101029] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101034] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1293660 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101503] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101517] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101522] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101527] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101533] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101538] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101543] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101548] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101552] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101560] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101566] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101571] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101575] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101581] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101586] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101591] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101596] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101601] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101606] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101611] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101616] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101621] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101626] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101630] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101635] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101640] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101645] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101650] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101655] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101660] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101664] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101669] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101673] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101679] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101683] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101688] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101699] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101704] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101708] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101713] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101718] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101723] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101728] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101733] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101737] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101742] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101746] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101751] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101755] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101761] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101765] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101770] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101775] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101779] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101784] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101788] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101793] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101798] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101803] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101807] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101812] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101817] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.101821] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x12939e0 is same with the state(6) to be set 00:22:19.282 [2024-10-17 17:43:27.110004] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.282 [2024-10-17 17:43:27.110035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110047] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110064] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110083] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110098] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11a2790 is same with the state(6) to be set 00:22:19.283 [2024-10-17 17:43:27.110127] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110147] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110163] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110180] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110195] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11a3310 is same with the state(6) to be set 00:22:19.283 [2024-10-17 17:43:27.110225] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110243] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110260] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110276] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110296] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd32c10 is same with the state(6) to be set 00:22:19.283 [2024-10-17 17:43:27.110317] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd26480 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.110346] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110365] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110382] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110399] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110415] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd3c4b0 is same with the state(6) to be set 00:22:19.283 [2024-10-17 17:43:27.110439] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110458] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110476] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110493] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110508] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd359a0 is same with the state(6) to be set 00:22:19.283 [2024-10-17 17:43:27.110534] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110553] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110569] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110589] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:19.283 [2024-10-17 17:43:27.110596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.110605] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc53610 is same with the state(6) to be set 00:22:19.283 [2024-10-17 17:43:27.112164] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6] resetting controller 00:22:19.283 [2024-10-17 17:43:27.112195] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc53610 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112236] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1165950 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112256] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1178a30 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112277] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11a2790 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112293] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11a3310 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112307] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd32c10 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112330] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd3c4b0 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112348] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd359a0 (9): Bad file descriptor 00:22:19.283 [2024-10-17 17:43:27.112384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112575] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.283 [2024-10-17 17:43:27.112618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.283 [2024-10-17 17:43:27.112632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.112986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.112994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113353] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.284 [2024-10-17 17:43:27.113533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.284 [2024-10-17 17:43:27.113546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.113769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.113777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.114739] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf3feb0 is same with the state(6) to be set 00:22:19.285 [2024-10-17 17:43:27.114790] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0xf3feb0 was disconnected and freed. reset controller. 00:22:19.285 [2024-10-17 17:43:27.114801] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.285 [2024-10-17 17:43:27.114850] nvme_tcp.c:1252:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:22:19.285 [2024-10-17 17:43:27.115593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.285 [2024-10-17 17:43:27.115617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd23800 with addr=10.0.0.2, port=4420 00:22:19.285 [2024-10-17 17:43:27.115627] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd23800 is same with the state(6) to be set 00:22:19.285 [2024-10-17 17:43:27.115734] nvme_tcp.c:1252:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:22:19.285 [2024-10-17 17:43:27.116123] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:22:19.285 [2024-10-17 17:43:27.116515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.285 [2024-10-17 17:43:27.116532] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc53610 with addr=10.0.0.2, port=4420 00:22:19.285 [2024-10-17 17:43:27.116541] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc53610 is same with the state(6) to be set 00:22:19.285 [2024-10-17 17:43:27.116552] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd23800 (9): Bad file descriptor 00:22:19.285 [2024-10-17 17:43:27.116874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.116887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.116900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.116909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.116921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.116929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.116940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.116948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.116959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.116968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.116978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.116987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.116998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:32768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:32896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:33024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:33152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.285 [2024-10-17 17:43:27.117371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.285 [2024-10-17 17:43:27.117380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.117983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.117990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.118001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.118010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.118021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.118029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.118040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.118049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.118059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.118068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.118078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.118087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.118098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.286 [2024-10-17 17:43:27.118108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.286 [2024-10-17 17:43:27.118118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118136] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11e0e70 is same with the state(6) to be set 00:22:19.287 [2024-10-17 17:43:27.118180] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x11e0e70 was disconnected and freed. reset controller. 00:22:19.287 [2024-10-17 17:43:27.118242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.118981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.118992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.119001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.119011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.119020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.119030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.287 [2024-10-17 17:43:27.119039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.287 [2024-10-17 17:43:27.119051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.119511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.119521] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x113efe0 is same with the state(6) to be set 00:22:19.288 [2024-10-17 17:43:27.119563] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x113efe0 was disconnected and freed. reset controller. 00:22:19.288 [2024-10-17 17:43:27.120100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.288 [2024-10-17 17:43:27.120142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd26480 with addr=10.0.0.2, port=4420 00:22:19.288 [2024-10-17 17:43:27.120155] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd26480 is same with the state(6) to be set 00:22:19.288 [2024-10-17 17:43:27.120172] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc53610 (9): Bad file descriptor 00:22:19.288 [2024-10-17 17:43:27.120185] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4] Ctrlr is in error state 00:22:19.288 [2024-10-17 17:43:27.120193] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4] controller reinitialization failed 00:22:19.288 [2024-10-17 17:43:27.120203] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4] in failed state. 00:22:19.288 [2024-10-17 17:43:27.123062] nvme_tcp.c:1252:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:22:19.288 [2024-10-17 17:43:27.123122] nvme_tcp.c:1252:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:22:19.288 [2024-10-17 17:43:27.123144] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.288 [2024-10-17 17:43:27.123156] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3] resetting controller 00:22:19.288 [2024-10-17 17:43:27.123171] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7] resetting controller 00:22:19.288 [2024-10-17 17:43:27.123202] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd26480 (9): Bad file descriptor 00:22:19.288 [2024-10-17 17:43:27.123214] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6] Ctrlr is in error state 00:22:19.288 [2024-10-17 17:43:27.123222] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6] controller reinitialization failed 00:22:19.288 [2024-10-17 17:43:27.123232] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6] in failed state. 00:22:19.288 [2024-10-17 17:43:27.123359] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.288 [2024-10-17 17:43:27.123684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.288 [2024-10-17 17:43:27.123709] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd3c4b0 with addr=10.0.0.2, port=4420 00:22:19.288 [2024-10-17 17:43:27.123718] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd3c4b0 is same with the state(6) to be set 00:22:19.288 [2024-10-17 17:43:27.124168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.288 [2024-10-17 17:43:27.124210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd32c10 with addr=10.0.0.2, port=4420 00:22:19.288 [2024-10-17 17:43:27.124222] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd32c10 is same with the state(6) to be set 00:22:19.288 [2024-10-17 17:43:27.124234] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:22:19.288 [2024-10-17 17:43:27.124242] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:22:19.288 [2024-10-17 17:43:27.124251] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:22:19.288 [2024-10-17 17:43:27.124305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.124318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.124336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.124346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.124357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.124366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.124377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.124386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.124397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.124407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.124418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.288 [2024-10-17 17:43:27.124427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.288 [2024-10-17 17:43:27.124438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.124980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.124991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.289 [2024-10-17 17:43:27.125282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.289 [2024-10-17 17:43:27.125293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.125605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.125615] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf41120 is same with the state(6) to be set 00:22:19.290 [2024-10-17 17:43:27.127174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.290 [2024-10-17 17:43:27.127464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.290 [2024-10-17 17:43:27.127474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127575] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.127990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.127997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.291 [2024-10-17 17:43:27.128230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.291 [2024-10-17 17:43:27.128238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.128253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.128261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.128272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.128280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.128290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.128298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.128308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.128316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.128327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.128335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.128345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.128353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.128362] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x113c670 is same with the state(6) to be set 00:22:19.292 [2024-10-17 17:43:27.129887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.129901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.129913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.129921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.129931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.129938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.129947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.129955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.129965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.129972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.129982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.129990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.292 [2024-10-17 17:43:27.130497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.292 [2024-10-17 17:43:27.130507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.130982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.130992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.131000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.131011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.131018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.131029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.131036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.131046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.131054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.131063] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1140560 is same with the state(6) to be set 00:22:19.293 [2024-10-17 17:43:27.132333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132383] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.293 [2024-10-17 17:43:27.132558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.293 [2024-10-17 17:43:27.132569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.132985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.132995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.294 [2024-10-17 17:43:27.133250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.294 [2024-10-17 17:43:27.133258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.133520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.133528] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11419a0 is same with the state(6) to be set 00:22:19.295 [2024-10-17 17:43:27.134800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.134983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.134993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135296] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.295 [2024-10-17 17:43:27.135315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.295 [2024-10-17 17:43:27.135325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135670] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.135986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:19.296 [2024-10-17 17:43:27.135994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:19.296 [2024-10-17 17:43:27.136002] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1142f20 is same with the state(6) to be set 00:22:19.296 [2024-10-17 17:43:27.138584] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4] resetting controller 00:22:19.296 [2024-10-17 17:43:27.138623] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.296 [2024-10-17 17:43:27.138639] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2] resetting controller 00:22:19.296 [2024-10-17 17:43:27.138649] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5] resetting controller 00:22:19.296 [2024-10-17 17:43:27.138659] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode8] resetting controller 00:22:19.296 [2024-10-17 17:43:27.138716] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd3c4b0 (9): Bad file descriptor 00:22:19.296 [2024-10-17 17:43:27.138731] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd32c10 (9): Bad file descriptor 00:22:19.296 [2024-10-17 17:43:27.138784] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.296 [2024-10-17 17:43:27.138799] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.296 [2024-10-17 17:43:27.138814] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.296 [2024-10-17 17:43:27.138826] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.296 [2024-10-17 17:43:27.138905] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode9] resetting controller 00:22:19.296 task offset: 26240 on job bdev=Nvme4n1 fails 00:22:19.297 00:22:19.297 Latency(us) 00:22:19.297 [2024-10-17T15:43:27.216Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:19.297 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme1n1 ended in about 0.95 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme1n1 : 0.95 201.21 12.58 67.07 0.00 235381.97 18677.76 246415.36 00:22:19.297 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme2n1 ended in about 0.97 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme2n1 : 0.97 198.69 12.42 66.23 0.00 233775.57 27525.12 237677.23 00:22:19.297 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme3n1 ended in about 0.96 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme3n1 : 0.96 203.95 12.75 66.60 0.00 224169.67 24466.77 228939.09 00:22:19.297 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme4n1 ended in about 0.93 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme4n1 : 0.93 205.50 12.84 68.50 0.00 216276.96 5270.19 244667.73 00:22:19.297 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme5n1 ended in about 0.97 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme5n1 : 0.97 132.09 8.26 66.04 0.00 293741.51 15291.73 255153.49 00:22:19.297 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme6n1 ended in about 0.95 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme6n1 : 0.95 201.78 12.61 67.26 0.00 211129.81 17803.95 249910.61 00:22:19.297 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme7n1 ended in about 0.96 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme7n1 : 0.96 140.27 8.77 66.50 0.00 269074.13 16711.68 272629.76 00:22:19.297 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme8n1 ended in about 0.97 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme8n1 : 0.97 197.58 12.35 65.86 0.00 206737.07 14964.05 251658.24 00:22:19.297 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme9n1 ended in about 0.97 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme9n1 : 0.97 131.39 8.21 65.69 0.00 270320.64 17694.72 267386.88 00:22:19.297 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:22:19.297 Job: Nvme10n1 ended in about 0.98 seconds with error 00:22:19.297 Verification LBA range: start 0x0 length 0x400 00:22:19.297 Nvme10n1 : 0.98 131.06 8.19 65.53 0.00 264949.76 19660.80 251658.24 00:22:19.297 [2024-10-17T15:43:27.216Z] =================================================================================================================== 00:22:19.297 [2024-10-17T15:43:27.216Z] Total : 1743.51 108.97 665.28 0.00 239069.25 5270.19 272629.76 00:22:19.297 [2024-10-17 17:43:27.165381] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:22:19.297 [2024-10-17 17:43:27.165431] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10] resetting controller 00:22:19.297 [2024-10-17 17:43:27.165863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.165885] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd23800 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.165897] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd23800 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.166089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.166101] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd359a0 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.166110] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd359a0 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.166389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.166401] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1165950 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.166408] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1165950 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.166735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.166747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1178a30 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.166754] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1178a30 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.166763] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3] Ctrlr is in error state 00:22:19.297 [2024-10-17 17:43:27.166770] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3] controller reinitialization failed 00:22:19.297 [2024-10-17 17:43:27.166779] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3] in failed state. 00:22:19.297 [2024-10-17 17:43:27.166792] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7] Ctrlr is in error state 00:22:19.297 [2024-10-17 17:43:27.166799] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7] controller reinitialization failed 00:22:19.297 [2024-10-17 17:43:27.166806] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7] in failed state. 00:22:19.297 [2024-10-17 17:43:27.168162] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6] resetting controller 00:22:19.297 [2024-10-17 17:43:27.168179] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:22:19.297 [2024-10-17 17:43:27.168189] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.297 [2024-10-17 17:43:27.168196] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.297 [2024-10-17 17:43:27.168562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.168578] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x11a3310 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.168592] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11a3310 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.168911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.168923] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x11a2790 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.168931] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11a2790 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.168944] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd23800 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.168955] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd359a0 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.168965] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1165950 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.168975] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1178a30 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.169016] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.297 [2024-10-17 17:43:27.169030] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.297 [2024-10-17 17:43:27.169042] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.297 [2024-10-17 17:43:27.169055] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:22:19.297 [2024-10-17 17:43:27.169573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.169591] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc53610 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.169599] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc53610 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.169792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.297 [2024-10-17 17:43:27.169804] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd26480 with addr=10.0.0.2, port=4420 00:22:19.297 [2024-10-17 17:43:27.169811] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd26480 is same with the state(6) to be set 00:22:19.297 [2024-10-17 17:43:27.169821] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11a3310 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.169833] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11a2790 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.169841] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4] Ctrlr is in error state 00:22:19.297 [2024-10-17 17:43:27.169848] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4] controller reinitialization failed 00:22:19.297 [2024-10-17 17:43:27.169856] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4] in failed state. 00:22:19.297 [2024-10-17 17:43:27.169868] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2] Ctrlr is in error state 00:22:19.297 [2024-10-17 17:43:27.169875] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2] controller reinitialization failed 00:22:19.297 [2024-10-17 17:43:27.169882] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2] in failed state. 00:22:19.297 [2024-10-17 17:43:27.169894] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5] Ctrlr is in error state 00:22:19.297 [2024-10-17 17:43:27.169900] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5] controller reinitialization failed 00:22:19.297 [2024-10-17 17:43:27.169907] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5] in failed state. 00:22:19.297 [2024-10-17 17:43:27.169921] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode8] Ctrlr is in error state 00:22:19.297 [2024-10-17 17:43:27.169928] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode8] controller reinitialization failed 00:22:19.297 [2024-10-17 17:43:27.169936] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode8] in failed state. 00:22:19.297 [2024-10-17 17:43:27.170000] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7] resetting controller 00:22:19.297 [2024-10-17 17:43:27.170011] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3] resetting controller 00:22:19.297 [2024-10-17 17:43:27.170020] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.297 [2024-10-17 17:43:27.170028] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.297 [2024-10-17 17:43:27.170035] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.297 [2024-10-17 17:43:27.170041] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.297 [2024-10-17 17:43:27.170062] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc53610 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.170073] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd26480 (9): Bad file descriptor 00:22:19.297 [2024-10-17 17:43:27.170081] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode9] Ctrlr is in error state 00:22:19.297 [2024-10-17 17:43:27.170088] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode9] controller reinitialization failed 00:22:19.298 [2024-10-17 17:43:27.170095] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode9] in failed state. 00:22:19.298 [2024-10-17 17:43:27.170105] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10] Ctrlr is in error state 00:22:19.298 [2024-10-17 17:43:27.170112] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10] controller reinitialization failed 00:22:19.298 [2024-10-17 17:43:27.170119] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10] in failed state. 00:22:19.298 [2024-10-17 17:43:27.170150] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.298 [2024-10-17 17:43:27.170158] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.298 [2024-10-17 17:43:27.170464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.298 [2024-10-17 17:43:27.170478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd32c10 with addr=10.0.0.2, port=4420 00:22:19.298 [2024-10-17 17:43:27.170486] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd32c10 is same with the state(6) to be set 00:22:19.298 [2024-10-17 17:43:27.170809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:19.298 [2024-10-17 17:43:27.170820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xd3c4b0 with addr=10.0.0.2, port=4420 00:22:19.298 [2024-10-17 17:43:27.170828] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xd3c4b0 is same with the state(6) to be set 00:22:19.298 [2024-10-17 17:43:27.170835] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6] Ctrlr is in error state 00:22:19.298 [2024-10-17 17:43:27.170842] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6] controller reinitialization failed 00:22:19.298 [2024-10-17 17:43:27.170851] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6] in failed state. 00:22:19.298 [2024-10-17 17:43:27.170861] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:22:19.298 [2024-10-17 17:43:27.170867] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:22:19.298 [2024-10-17 17:43:27.170875] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:22:19.298 [2024-10-17 17:43:27.170907] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.298 [2024-10-17 17:43:27.170915] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.298 [2024-10-17 17:43:27.170923] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd32c10 (9): Bad file descriptor 00:22:19.298 [2024-10-17 17:43:27.170933] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd3c4b0 (9): Bad file descriptor 00:22:19.298 [2024-10-17 17:43:27.170961] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7] Ctrlr is in error state 00:22:19.298 [2024-10-17 17:43:27.170970] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7] controller reinitialization failed 00:22:19.298 [2024-10-17 17:43:27.170977] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7] in failed state. 00:22:19.298 [2024-10-17 17:43:27.170987] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3] Ctrlr is in error state 00:22:19.298 [2024-10-17 17:43:27.170994] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3] controller reinitialization failed 00:22:19.298 [2024-10-17 17:43:27.171001] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3] in failed state. 00:22:19.298 [2024-10-17 17:43:27.171031] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.298 [2024-10-17 17:43:27.171039] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:19.569 17:43:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@137 -- # sleep 1 00:22:20.513 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@138 -- # NOT wait 116296 00:22:20.513 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@650 -- # local es=0 00:22:20.513 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@652 -- # valid_exec_arg wait 116296 00:22:20.513 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@638 -- # local arg=wait 00:22:20.513 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@642 -- # type -t wait 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@653 -- # wait 116296 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@653 -- # es=255 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@662 -- # es=127 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@663 -- # case "$es" in 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@670 -- # es=1 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@140 -- # stoptarget 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@46 -- # nvmftestfini 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@121 -- # sync 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@124 -- # set +e 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:20.514 rmmod nvme_tcp 00:22:20.514 rmmod nvme_fabrics 00:22:20.514 rmmod nvme_keyring 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@128 -- # set -e 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@129 -- # return 0 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@515 -- # '[' -n 115930 ']' 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@516 -- # killprocess 115930 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@950 -- # '[' -z 115930 ']' 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # kill -0 115930 00:22:20.514 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (115930) - No such process 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@977 -- # echo 'Process with pid 115930 is not found' 00:22:20.514 Process with pid 115930 is not found 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@297 -- # iptr 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@789 -- # iptables-save 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:20.514 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@789 -- # iptables-restore 00:22:20.775 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:20.775 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:20.775 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:20.775 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:20.775 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:22.690 00:22:22.690 real 0m7.564s 00:22:22.690 user 0m17.966s 00:22:22.690 sys 0m1.241s 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:22:22.690 ************************************ 00:22:22.690 END TEST nvmf_shutdown_tc3 00:22:22.690 ************************************ 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ e810 == \e\8\1\0 ]] 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ tcp == \r\d\m\a ]] 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@167 -- # run_test nvmf_shutdown_tc4 nvmf_shutdown_tc4 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:22:22.690 ************************************ 00:22:22.690 START TEST nvmf_shutdown_tc4 00:22:22.690 ************************************ 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc4 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@145 -- # starttarget 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@16 -- # nvmftestinit 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@474 -- # prepare_net_devs 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@309 -- # xtrace_disable 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # pci_devs=() 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # net_devs=() 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # e810=() 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # local -ga e810 00:22:22.690 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # x722=() 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # local -ga x722 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # mlx=() 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # local -ga mlx 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:22.951 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:22.952 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:22.952 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:22.952 Found net devices under 0000:31:00.0: cvl_0_0 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:22.952 Found net devices under 0000:31:00.1: cvl_0_1 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # is_hw=yes 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:22.952 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:23.214 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:23.214 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.579 ms 00:22:23.214 00:22:23.214 --- 10.0.0.2 ping statistics --- 00:22:23.214 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:23.214 rtt min/avg/max/mdev = 0.579/0.579/0.579/0.000 ms 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:23.214 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:23.214 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.247 ms 00:22:23.214 00:22:23.214 --- 10.0.0.1 ping statistics --- 00:22:23.214 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:23.214 rtt min/avg/max/mdev = 0.247/0.247/0.247/0.000 ms 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@448 -- # return 0 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@507 -- # nvmfpid=117696 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@508 -- # waitforlisten 117696 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@831 -- # '[' -z 117696 ']' 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:23.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:23.214 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:23.214 [2024-10-17 17:43:31.018057] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:23.214 [2024-10-17 17:43:31.018112] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:23.214 [2024-10-17 17:43:31.106611] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:23.475 [2024-10-17 17:43:31.147766] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:23.475 [2024-10-17 17:43:31.147806] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:23.475 [2024-10-17 17:43:31.147812] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:23.475 [2024-10-17 17:43:31.147818] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:23.475 [2024-10-17 17:43:31.147823] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:23.475 [2024-10-17 17:43:31.149329] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:23.475 [2024-10-17 17:43:31.149485] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:23.475 [2024-10-17 17:43:31.149641] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:23.475 [2024-10-17 17:43:31.149643] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@864 -- # return 0 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:24.046 [2024-10-17 17:43:31.870712] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@36 -- # rpc_cmd 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.046 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:24.046 Malloc1 00:22:24.307 [2024-10-17 17:43:31.981556] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:24.307 Malloc2 00:22:24.307 Malloc3 00:22:24.307 Malloc4 00:22:24.307 Malloc5 00:22:24.307 Malloc6 00:22:24.307 Malloc7 00:22:24.567 Malloc8 00:22:24.567 Malloc9 00:22:24.567 Malloc10 00:22:24.567 17:43:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.567 17:43:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:22:24.567 17:43:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:24.567 17:43:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:24.567 17:43:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@149 -- # perfpid=117937 00:22:24.567 17:43:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@150 -- # sleep 5 00:22:24.567 17:43:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@148 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 45056 -O 4096 -w randwrite -t 20 -r 'trtype:tcp adrfam:IPV4 traddr:10.0.0.2 trsvcid:4420' -P 4 00:22:24.567 [2024-10-17 17:43:32.454575] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@152 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@155 -- # killprocess 117696 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@950 -- # '[' -z 117696 ']' 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # kill -0 117696 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@955 -- # uname 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 117696 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 117696' 00:22:29.858 killing process with pid 117696 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@969 -- # kill 117696 00:22:29.858 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@974 -- # wait 117696 00:22:29.858 [2024-10-17 17:43:37.457439] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093af0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457485] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093af0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457491] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093af0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457497] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093af0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457502] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093af0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457506] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093af0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457511] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093af0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457828] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093fc0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457869] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093fc0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457876] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093fc0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.457882] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093fc0 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458457] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458483] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458489] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458494] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458499] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458504] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458509] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458513] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2094490 is same with the state(6) to be set 00:22:29.858 [2024-10-17 17:43:37.458940] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093620 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.458967] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093620 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.458974] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2093620 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462193] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1950 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462213] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1950 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462218] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1950 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462536] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1e20 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462553] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1e20 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462558] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1e20 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462564] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1e20 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462570] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1e20 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462575] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1e20 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462580] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c1e20 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462748] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462766] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462771] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462788] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462793] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462798] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462803] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462808] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.462813] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c22f0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463070] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1e53ce0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463089] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1e53ce0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463095] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1e53ce0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463100] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1e53ce0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463106] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1e53ce0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463793] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463810] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463816] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463821] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463826] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463831] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463836] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463841] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463845] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463850] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.463855] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2095300 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.464226] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20957d0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.464242] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20957d0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.464247] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20957d0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.464253] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20957d0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.464259] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20957d0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.466814] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c2cb0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.466834] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c2cb0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.466839] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c2cb0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.466844] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c2cb0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.466849] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c2cb0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.466854] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c2cb0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.466858] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c2cb0 is same with the state(6) to be set 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 [2024-10-17 17:43:37.467202] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c31a0 is same with starting I/O failed: -6 00:22:29.859 the state(6) to be set 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 [2024-10-17 17:43:37.467221] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c31a0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467226] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c31a0 is same with the state(6) to be set 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 [2024-10-17 17:43:37.467490] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3690 is same with the state(6) to be set 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 [2024-10-17 17:43:37.467507] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3690 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467513] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3690 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467519] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3690 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467524] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3690 is same with the state(6) to be set 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 [2024-10-17 17:43:37.467529] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3690 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467535] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3690 is same with the state(6) to be set 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 starting I/O failed: -6 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 Write completed with error (sct=0, sc=8) 00:22:29.859 [2024-10-17 17:43:37.467815] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c27c0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467832] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c27c0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467838] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c27c0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467844] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c27c0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467849] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c27c0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467855] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c27c0 is same with the state(6) to be set 00:22:29.859 [2024-10-17 17:43:37.467830] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.859 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.468044] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2011120 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468058] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2011120 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468063] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2011120 is same with the state(6) to be set 00:22:29.860 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.468068] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2011120 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468073] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2011120 is same with the state(6) to be set 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 [2024-10-17 17:43:37.468282] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20115f0 is same with Write completed with error (sct=0, sc=8) 00:22:29.860 the state(6) to be set 00:22:29.860 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.468296] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20115f0 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468302] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20115f0 is same with the state(6) to be set 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 [2024-10-17 17:43:37.468307] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20115f0 is same with the state(6) to be set 00:22:29.860 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.468312] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20115f0 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468317] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20115f0 is same with the state(6) to be set 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.468555] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3a30 is same with the state(6) to be set 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 [2024-10-17 17:43:37.468570] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3a30 is same with the state(6) to be set 00:22:29.860 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.468576] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3a30 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468582] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3a30 is same with the state(6) to be set 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 [2024-10-17 17:43:37.468588] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3a30 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468593] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3a30 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468598] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20c3a30 is same with the state(6) to be set 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.468715] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.860 [2024-10-17 17:43:37.468787] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2010c50 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468798] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2010c50 is same with the state(6) to be set 00:22:29.860 [2024-10-17 17:43:37.468803] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2010c50 is same with the state(6) to be set 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 [2024-10-17 17:43:37.469623] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.860 Write completed with error (sct=0, sc=8) 00:22:29.860 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 [2024-10-17 17:43:37.471037] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.861 NVMe io qpair process completion error 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 [2024-10-17 17:43:37.472200] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 [2024-10-17 17:43:37.473029] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 Write completed with error (sct=0, sc=8) 00:22:29.861 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 [2024-10-17 17:43:37.473970] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 [2024-10-17 17:43:37.475634] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.862 NVMe io qpair process completion error 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 starting I/O failed: -6 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.862 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 [2024-10-17 17:43:37.476933] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 [2024-10-17 17:43:37.477754] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 [2024-10-17 17:43:37.478684] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.863 Write completed with error (sct=0, sc=8) 00:22:29.863 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 [2024-10-17 17:43:37.482158] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.864 NVMe io qpair process completion error 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 [2024-10-17 17:43:37.483646] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.864 starting I/O failed: -6 00:22:29.864 starting I/O failed: -6 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 [2024-10-17 17:43:37.484593] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.864 Write completed with error (sct=0, sc=8) 00:22:29.864 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 [2024-10-17 17:43:37.485499] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 [2024-10-17 17:43:37.486949] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.865 NVMe io qpair process completion error 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 starting I/O failed: -6 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.865 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 [2024-10-17 17:43:37.488080] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 [2024-10-17 17:43:37.488904] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 [2024-10-17 17:43:37.489849] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.866 starting I/O failed: -6 00:22:29.866 starting I/O failed: -6 00:22:29.866 starting I/O failed: -6 00:22:29.866 starting I/O failed: -6 00:22:29.866 starting I/O failed: -6 00:22:29.866 starting I/O failed: -6 00:22:29.866 starting I/O failed: -6 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.866 starting I/O failed: -6 00:22:29.866 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 [2024-10-17 17:43:37.492647] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.867 NVMe io qpair process completion error 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 [2024-10-17 17:43:37.493909] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 [2024-10-17 17:43:37.494735] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 starting I/O failed: -6 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.867 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 [2024-10-17 17:43:37.495650] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 [2024-10-17 17:43:37.497264] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.868 NVMe io qpair process completion error 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 Write completed with error (sct=0, sc=8) 00:22:29.868 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 [2024-10-17 17:43:37.498584] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 [2024-10-17 17:43:37.499434] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 [2024-10-17 17:43:37.500368] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.869 Write completed with error (sct=0, sc=8) 00:22:29.869 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 [2024-10-17 17:43:37.503744] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.870 NVMe io qpair process completion error 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 [2024-10-17 17:43:37.505014] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 [2024-10-17 17:43:37.505851] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.870 starting I/O failed: -6 00:22:29.870 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 [2024-10-17 17:43:37.506771] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 [2024-10-17 17:43:37.508221] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.871 NVMe io qpair process completion error 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 starting I/O failed: -6 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.871 Write completed with error (sct=0, sc=8) 00:22:29.872 [2024-10-17 17:43:37.509370] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 [2024-10-17 17:43:37.510185] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 [2024-10-17 17:43:37.511113] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.872 Write completed with error (sct=0, sc=8) 00:22:29.872 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 [2024-10-17 17:43:37.512548] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.873 NVMe io qpair process completion error 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 [2024-10-17 17:43:37.513801] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 [2024-10-17 17:43:37.514603] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.873 Write completed with error (sct=0, sc=8) 00:22:29.873 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 [2024-10-17 17:43:37.515544] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 Write completed with error (sct=0, sc=8) 00:22:29.874 starting I/O failed: -6 00:22:29.874 [2024-10-17 17:43:37.519761] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:22:29.874 NVMe io qpair process completion error 00:22:29.874 Initializing NVMe Controllers 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode3 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode7 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode8 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode6 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode5 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode9 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode2 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode10 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode4 00:22:29.874 Controller IO queue size 128, less than required. 00:22:29.874 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:22:29.874 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 with lcore 0 00:22:29.874 Initialization complete. Launching workers. 00:22:29.875 ======================================================== 00:22:29.875 Latency(us) 00:22:29.875 Device Information : IOPS MiB/s Average min max 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 from core 0: 1887.32 81.10 67839.17 852.91 132397.44 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 from core 0: 1896.80 81.50 67517.81 816.44 126272.69 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 from core 0: 1907.55 81.97 67156.76 850.25 118784.34 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 from core 0: 1874.05 80.53 67652.87 643.66 124043.15 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 from core 0: 1892.59 81.32 67008.51 684.15 125754.75 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 from core 0: 1843.29 79.20 68824.22 678.87 118876.51 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 from core 0: 1867.52 80.24 67984.25 680.11 123752.60 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 from core 0: 1881.21 80.83 67511.72 679.54 123557.89 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1883.95 80.95 67448.25 856.61 123442.28 00:22:29.875 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 from core 0: 1896.17 81.48 67037.21 664.16 123362.16 00:22:29.875 ======================================================== 00:22:29.875 Total : 18830.47 809.12 67593.88 643.66 132397.44 00:22:29.875 00:22:29.875 [2024-10-17 17:43:37.522492] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x19575b0 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522535] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x19584e0 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522565] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1958810 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522597] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x19581b0 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522626] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1957fd0 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522657] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1958b40 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522783] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1957280 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522821] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1959760 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522849] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1959430 is same with the state(6) to be set 00:22:29.875 [2024-10-17 17:43:37.522880] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x19578e0 is same with the state(6) to be set 00:22:29.875 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:22:29.875 17:43:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@156 -- # sleep 1 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@158 -- # NOT wait 117937 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@650 -- # local es=0 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@652 -- # valid_exec_arg wait 117937 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@638 -- # local arg=wait 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@642 -- # type -t wait 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@653 -- # wait 117937 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@653 -- # es=1 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@159 -- # stoptarget 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@46 -- # nvmftestfini 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@121 -- # sync 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@124 -- # set +e 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:30.818 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:30.818 rmmod nvme_tcp 00:22:31.079 rmmod nvme_fabrics 00:22:31.079 rmmod nvme_keyring 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@128 -- # set -e 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@129 -- # return 0 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@515 -- # '[' -n 117696 ']' 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@516 -- # killprocess 117696 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@950 -- # '[' -z 117696 ']' 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # kill -0 117696 00:22:31.079 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (117696) - No such process 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@977 -- # echo 'Process with pid 117696 is not found' 00:22:31.079 Process with pid 117696 is not found 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@297 -- # iptr 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@789 -- # iptables-save 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@789 -- # iptables-restore 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:31.079 17:43:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:33.059 17:43:40 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:33.059 00:22:33.059 real 0m10.277s 00:22:33.059 user 0m27.751s 00:22:33.059 sys 0m4.099s 00:22:33.059 17:43:40 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:33.059 17:43:40 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:22:33.059 ************************************ 00:22:33.059 END TEST nvmf_shutdown_tc4 00:22:33.059 ************************************ 00:22:33.059 17:43:40 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@170 -- # trap - SIGINT SIGTERM EXIT 00:22:33.059 00:22:33.059 real 0m42.947s 00:22:33.059 user 1m43.597s 00:22:33.059 sys 0m13.647s 00:22:33.059 17:43:40 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:33.059 17:43:40 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:22:33.059 ************************************ 00:22:33.059 END TEST nvmf_shutdown 00:22:33.059 ************************************ 00:22:33.344 17:43:40 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:22:33.344 00:22:33.344 real 12m50.202s 00:22:33.344 user 27m8.616s 00:22:33.344 sys 3m52.235s 00:22:33.344 17:43:40 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:33.344 17:43:40 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:22:33.344 ************************************ 00:22:33.344 END TEST nvmf_target_extra 00:22:33.344 ************************************ 00:22:33.344 17:43:40 nvmf_tcp -- nvmf/nvmf.sh@16 -- # run_test nvmf_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:22:33.344 17:43:40 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:22:33.344 17:43:41 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:33.344 17:43:41 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:22:33.344 ************************************ 00:22:33.344 START TEST nvmf_host 00:22:33.344 ************************************ 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:22:33.344 * Looking for test storage... 00:22:33.344 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1691 -- # lcov --version 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # IFS=.-: 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # read -ra ver1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # IFS=.-: 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # read -ra ver2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@338 -- # local 'op=<' 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@340 -- # ver1_l=2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@341 -- # ver2_l=1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@344 -- # case "$op" in 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@345 -- # : 1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # decimal 1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # ver1[v]=1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # decimal 2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # ver2[v]=2 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # return 0 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:22:33.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.344 --rc genhtml_branch_coverage=1 00:22:33.344 --rc genhtml_function_coverage=1 00:22:33.344 --rc genhtml_legend=1 00:22:33.344 --rc geninfo_all_blocks=1 00:22:33.344 --rc geninfo_unexecuted_blocks=1 00:22:33.344 00:22:33.344 ' 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:22:33.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.344 --rc genhtml_branch_coverage=1 00:22:33.344 --rc genhtml_function_coverage=1 00:22:33.344 --rc genhtml_legend=1 00:22:33.344 --rc geninfo_all_blocks=1 00:22:33.344 --rc geninfo_unexecuted_blocks=1 00:22:33.344 00:22:33.344 ' 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:22:33.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.344 --rc genhtml_branch_coverage=1 00:22:33.344 --rc genhtml_function_coverage=1 00:22:33.344 --rc genhtml_legend=1 00:22:33.344 --rc geninfo_all_blocks=1 00:22:33.344 --rc geninfo_unexecuted_blocks=1 00:22:33.344 00:22:33.344 ' 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:22:33.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.344 --rc genhtml_branch_coverage=1 00:22:33.344 --rc genhtml_function_coverage=1 00:22:33.344 --rc genhtml_legend=1 00:22:33.344 --rc geninfo_all_blocks=1 00:22:33.344 --rc geninfo_unexecuted_blocks=1 00:22:33.344 00:22:33.344 ' 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # uname -s 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:33.344 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@15 -- # shopt -s extglob 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- paths/export.sh@5 -- # export PATH 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@51 -- # : 0 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:22:33.612 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@13 -- # TEST_ARGS=("$@") 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@15 -- # [[ 0 -eq 0 ]] 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@16 -- # run_test nvmf_multicontroller /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:22:33.612 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:22:33.613 ************************************ 00:22:33.613 START TEST nvmf_multicontroller 00:22:33.613 ************************************ 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:22:33.613 * Looking for test storage... 00:22:33.613 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1691 -- # lcov --version 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@333 -- # local ver1 ver1_l 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@334 -- # local ver2 ver2_l 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # IFS=.-: 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # read -ra ver1 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # IFS=.-: 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # read -ra ver2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@338 -- # local 'op=<' 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@340 -- # ver1_l=2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@341 -- # ver2_l=1 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@344 -- # case "$op" in 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@345 -- # : 1 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v = 0 )) 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # decimal 1 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=1 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 1 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # ver1[v]=1 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # decimal 2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # ver2[v]=2 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # return 0 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:22:33.613 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.613 --rc genhtml_branch_coverage=1 00:22:33.613 --rc genhtml_function_coverage=1 00:22:33.613 --rc genhtml_legend=1 00:22:33.613 --rc geninfo_all_blocks=1 00:22:33.613 --rc geninfo_unexecuted_blocks=1 00:22:33.613 00:22:33.613 ' 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:22:33.613 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.613 --rc genhtml_branch_coverage=1 00:22:33.613 --rc genhtml_function_coverage=1 00:22:33.613 --rc genhtml_legend=1 00:22:33.613 --rc geninfo_all_blocks=1 00:22:33.613 --rc geninfo_unexecuted_blocks=1 00:22:33.613 00:22:33.613 ' 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:22:33.613 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.613 --rc genhtml_branch_coverage=1 00:22:33.613 --rc genhtml_function_coverage=1 00:22:33.613 --rc genhtml_legend=1 00:22:33.613 --rc geninfo_all_blocks=1 00:22:33.613 --rc geninfo_unexecuted_blocks=1 00:22:33.613 00:22:33.613 ' 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:22:33.613 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:33.613 --rc genhtml_branch_coverage=1 00:22:33.613 --rc genhtml_function_coverage=1 00:22:33.613 --rc genhtml_legend=1 00:22:33.613 --rc geninfo_all_blocks=1 00:22:33.613 --rc geninfo_unexecuted_blocks=1 00:22:33.613 00:22:33.613 ' 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:33.613 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # uname -s 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@15 -- # shopt -s extglob 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@5 -- # export PATH 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@51 -- # : 0 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:22:33.934 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@55 -- # have_pci_nics=0 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@11 -- # MALLOC_BDEV_SIZE=64 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@13 -- # NVMF_HOST_FIRST_PORT=60000 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@14 -- # NVMF_HOST_SECOND_PORT=60001 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@18 -- # '[' tcp == rdma ']' 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@23 -- # nvmftestinit 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@474 -- # prepare_net_devs 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@436 -- # local -g is_hw=no 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@438 -- # remove_spdk_ns 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@309 -- # xtrace_disable 00:22:33.934 17:43:41 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # pci_devs=() 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # net_devs=() 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # e810=() 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # local -ga e810 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # x722=() 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # local -ga x722 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # mlx=() 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # local -ga mlx 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:42.099 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:42.099 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:42.099 Found net devices under 0000:31:00.0: cvl_0_0 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:42.099 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:42.100 Found net devices under 0000:31:00.1: cvl_0_1 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # is_hw=yes 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:42.100 17:43:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:42.100 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:42.100 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.630 ms 00:22:42.100 00:22:42.100 --- 10.0.0.2 ping statistics --- 00:22:42.100 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:42.100 rtt min/avg/max/mdev = 0.630/0.630/0.630/0.000 ms 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:42.100 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:42.100 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.291 ms 00:22:42.100 00:22:42.100 --- 10.0.0.1 ping statistics --- 00:22:42.100 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:42.100 rtt min/avg/max/mdev = 0.291/0.291/0.291/0.000 ms 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@448 -- # return 0 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@25 -- # nvmfappstart -m 0xE 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@507 -- # nvmfpid=123590 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@508 -- # waitforlisten 123590 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@831 -- # '[' -z 123590 ']' 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:42.100 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:42.100 17:43:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.100 [2024-10-17 17:43:49.350021] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:42.100 [2024-10-17 17:43:49.350113] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:42.100 [2024-10-17 17:43:49.440899] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:22:42.100 [2024-10-17 17:43:49.492892] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:42.100 [2024-10-17 17:43:49.492945] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:42.100 [2024-10-17 17:43:49.492954] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:42.100 [2024-10-17 17:43:49.492961] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:42.100 [2024-10-17 17:43:49.492968] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:42.100 [2024-10-17 17:43:49.495105] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:42.100 [2024-10-17 17:43:49.495266] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:42.100 [2024-10-17 17:43:49.495266] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # return 0 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@27 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.361 [2024-10-17 17:43:50.218383] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@29 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.361 Malloc0 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@30 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.361 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.622 [2024-10-17 17:43:50.291364] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.622 [2024-10-17 17:43:50.303192] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.622 Malloc1 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc1 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4421 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:42.622 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@44 -- # bdevperf_pid=123750 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@46 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; pap "$testdir/try.txt"; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w write -t 1 -f 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@47 -- # waitforlisten 123750 /var/tmp/bdevperf.sock 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@831 -- # '[' -z 123750 ']' 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:22:42.623 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:42.623 17:43:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # return 0 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@50 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:43.566 NVMe0n1 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # grep -c NVMe 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.566 1 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@60 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.566 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:43.827 request: 00:22:43.827 { 00:22:43.827 "name": "NVMe0", 00:22:43.827 "trtype": "tcp", 00:22:43.827 "traddr": "10.0.0.2", 00:22:43.827 "adrfam": "ipv4", 00:22:43.827 "trsvcid": "4420", 00:22:43.827 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:22:43.827 "hostnqn": "nqn.2021-09-7.io.spdk:00001", 00:22:43.827 "hostaddr": "10.0.0.1", 00:22:43.827 "prchk_reftag": false, 00:22:43.827 "prchk_guard": false, 00:22:43.827 "hdgst": false, 00:22:43.827 "ddgst": false, 00:22:43.827 "allow_unrecognized_csi": false, 00:22:43.827 "method": "bdev_nvme_attach_controller", 00:22:43.827 "req_id": 1 00:22:43.827 } 00:22:43.827 Got JSON-RPC error response 00:22:43.827 response: 00:22:43.827 { 00:22:43.827 "code": -114, 00:22:43.827 "message": "A controller named NVMe0 already exists with the specified network path" 00:22:43.827 } 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@65 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:43.827 request: 00:22:43.827 { 00:22:43.827 "name": "NVMe0", 00:22:43.827 "trtype": "tcp", 00:22:43.827 "traddr": "10.0.0.2", 00:22:43.827 "adrfam": "ipv4", 00:22:43.827 "trsvcid": "4420", 00:22:43.827 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:22:43.827 "hostaddr": "10.0.0.1", 00:22:43.827 "prchk_reftag": false, 00:22:43.827 "prchk_guard": false, 00:22:43.827 "hdgst": false, 00:22:43.827 "ddgst": false, 00:22:43.827 "allow_unrecognized_csi": false, 00:22:43.827 "method": "bdev_nvme_attach_controller", 00:22:43.827 "req_id": 1 00:22:43.827 } 00:22:43.827 Got JSON-RPC error response 00:22:43.827 response: 00:22:43.827 { 00:22:43.827 "code": -114, 00:22:43.827 "message": "A controller named NVMe0 already exists with the specified network path" 00:22:43.827 } 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@69 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:43.827 request: 00:22:43.827 { 00:22:43.827 "name": "NVMe0", 00:22:43.827 "trtype": "tcp", 00:22:43.827 "traddr": "10.0.0.2", 00:22:43.827 "adrfam": "ipv4", 00:22:43.827 "trsvcid": "4420", 00:22:43.827 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:22:43.827 "hostaddr": "10.0.0.1", 00:22:43.827 "prchk_reftag": false, 00:22:43.827 "prchk_guard": false, 00:22:43.827 "hdgst": false, 00:22:43.827 "ddgst": false, 00:22:43.827 "multipath": "disable", 00:22:43.827 "allow_unrecognized_csi": false, 00:22:43.827 "method": "bdev_nvme_attach_controller", 00:22:43.827 "req_id": 1 00:22:43.827 } 00:22:43.827 Got JSON-RPC error response 00:22:43.827 response: 00:22:43.827 { 00:22:43.827 "code": -114, 00:22:43.827 "message": "A controller named NVMe0 already exists and multipath is disabled" 00:22:43.827 } 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@74 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:43.827 request: 00:22:43.827 { 00:22:43.827 "name": "NVMe0", 00:22:43.827 "trtype": "tcp", 00:22:43.827 "traddr": "10.0.0.2", 00:22:43.827 "adrfam": "ipv4", 00:22:43.827 "trsvcid": "4420", 00:22:43.827 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:22:43.827 "hostaddr": "10.0.0.1", 00:22:43.827 "prchk_reftag": false, 00:22:43.827 "prchk_guard": false, 00:22:43.827 "hdgst": false, 00:22:43.827 "ddgst": false, 00:22:43.827 "multipath": "failover", 00:22:43.827 "allow_unrecognized_csi": false, 00:22:43.827 "method": "bdev_nvme_attach_controller", 00:22:43.827 "req_id": 1 00:22:43.827 } 00:22:43.827 Got JSON-RPC error response 00:22:43.827 response: 00:22:43.827 { 00:22:43.827 "code": -114, 00:22:43.827 "message": "A controller named NVMe0 already exists with the specified network path" 00:22:43.827 } 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@79 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.827 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:44.088 NVMe0n1 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@83 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@87 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe1 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:44.088 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # grep -c NVMe 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # '[' 2 '!=' 2 ']' 00:22:44.088 17:43:51 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:22:45.471 { 00:22:45.471 "results": [ 00:22:45.471 { 00:22:45.471 "job": "NVMe0n1", 00:22:45.471 "core_mask": "0x1", 00:22:45.471 "workload": "write", 00:22:45.471 "status": "finished", 00:22:45.471 "queue_depth": 128, 00:22:45.471 "io_size": 4096, 00:22:45.471 "runtime": 1.005098, 00:22:45.472 "iops": 26849.123170078936, 00:22:45.472 "mibps": 104.87938738312084, 00:22:45.472 "io_failed": 0, 00:22:45.472 "io_timeout": 0, 00:22:45.472 "avg_latency_us": 4756.600793497863, 00:22:45.472 "min_latency_us": 2116.266666666667, 00:22:45.472 "max_latency_us": 16930.133333333335 00:22:45.472 } 00:22:45.472 ], 00:22:45.472 "core_count": 1 00:22:45.472 } 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@98 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe1 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@100 -- # [[ -n '' ]] 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@116 -- # killprocess 123750 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@950 -- # '[' -z 123750 ']' 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # kill -0 123750 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # uname 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 123750 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@968 -- # echo 'killing process with pid 123750' 00:22:45.472 killing process with pid 123750 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@969 -- # kill 123750 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@974 -- # wait 123750 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@118 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@119 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@123 -- # pap /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1597 -- # read -r file 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1596 -- # find /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt -type f 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1596 -- # sort -u 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1598 -- # cat 00:22:45.472 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:22:45.472 [2024-10-17 17:43:50.434522] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:45.472 [2024-10-17 17:43:50.434591] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid123750 ] 00:22:45.472 [2024-10-17 17:43:50.517889] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:45.472 [2024-10-17 17:43:50.571912] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:45.472 [2024-10-17 17:43:51.959403] bdev.c:4701:bdev_name_add: *ERROR*: Bdev name 2c18359d-cff8-46dc-bccd-fe964be65b11 already exists 00:22:45.472 [2024-10-17 17:43:51.959447] bdev.c:7846:bdev_register: *ERROR*: Unable to add uuid:2c18359d-cff8-46dc-bccd-fe964be65b11 alias for bdev NVMe1n1 00:22:45.472 [2024-10-17 17:43:51.959457] bdev_nvme.c:4484:nvme_bdev_create: *ERROR*: spdk_bdev_register() failed 00:22:45.472 Running I/O for 1 seconds... 00:22:45.472 26808.00 IOPS, 104.72 MiB/s 00:22:45.472 Latency(us) 00:22:45.472 [2024-10-17T15:43:53.391Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:45.472 Job: NVMe0n1 (Core Mask 0x1, workload: write, depth: 128, IO size: 4096) 00:22:45.472 NVMe0n1 : 1.01 26849.12 104.88 0.00 0.00 4756.60 2116.27 16930.13 00:22:45.472 [2024-10-17T15:43:53.391Z] =================================================================================================================== 00:22:45.472 [2024-10-17T15:43:53.391Z] Total : 26849.12 104.88 0.00 0.00 4756.60 2116.27 16930.13 00:22:45.472 Received shutdown signal, test time was about 1.000000 seconds 00:22:45.472 00:22:45.472 Latency(us) 00:22:45.472 [2024-10-17T15:43:53.391Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:45.472 [2024-10-17T15:43:53.391Z] =================================================================================================================== 00:22:45.472 [2024-10-17T15:43:53.391Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:22:45.472 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1603 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1597 -- # read -r file 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@124 -- # nvmftestfini 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@121 -- # sync 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@124 -- # set +e 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:45.472 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:45.472 rmmod nvme_tcp 00:22:45.472 rmmod nvme_fabrics 00:22:45.732 rmmod nvme_keyring 00:22:45.732 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:45.732 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@128 -- # set -e 00:22:45.732 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@129 -- # return 0 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@515 -- # '[' -n 123590 ']' 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@516 -- # killprocess 123590 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@950 -- # '[' -z 123590 ']' 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # kill -0 123590 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # uname 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 123590 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@968 -- # echo 'killing process with pid 123590' 00:22:45.733 killing process with pid 123590 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@969 -- # kill 123590 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@974 -- # wait 123590 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@297 -- # iptr 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@789 -- # iptables-save 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@789 -- # iptables-restore 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:45.733 17:43:53 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:48.279 00:22:48.279 real 0m14.387s 00:22:48.279 user 0m17.781s 00:22:48.279 sys 0m6.765s 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:22:48.279 ************************************ 00:22:48.279 END TEST nvmf_multicontroller 00:22:48.279 ************************************ 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@17 -- # run_test nvmf_aer /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:22:48.279 ************************************ 00:22:48.279 START TEST nvmf_aer 00:22:48.279 ************************************ 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:22:48.279 * Looking for test storage... 00:22:48.279 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1691 -- # lcov --version 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@333 -- # local ver1 ver1_l 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@334 -- # local ver2 ver2_l 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # IFS=.-: 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # read -ra ver1 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # IFS=.-: 00:22:48.279 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # read -ra ver2 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@338 -- # local 'op=<' 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@340 -- # ver1_l=2 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@341 -- # ver2_l=1 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@344 -- # case "$op" in 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@345 -- # : 1 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v = 0 )) 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # decimal 1 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=1 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 1 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # ver1[v]=1 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # decimal 2 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=2 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 2 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # ver2[v]=2 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # return 0 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:22:48.280 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:48.280 --rc genhtml_branch_coverage=1 00:22:48.280 --rc genhtml_function_coverage=1 00:22:48.280 --rc genhtml_legend=1 00:22:48.280 --rc geninfo_all_blocks=1 00:22:48.280 --rc geninfo_unexecuted_blocks=1 00:22:48.280 00:22:48.280 ' 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:22:48.280 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:48.280 --rc genhtml_branch_coverage=1 00:22:48.280 --rc genhtml_function_coverage=1 00:22:48.280 --rc genhtml_legend=1 00:22:48.280 --rc geninfo_all_blocks=1 00:22:48.280 --rc geninfo_unexecuted_blocks=1 00:22:48.280 00:22:48.280 ' 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:22:48.280 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:48.280 --rc genhtml_branch_coverage=1 00:22:48.280 --rc genhtml_function_coverage=1 00:22:48.280 --rc genhtml_legend=1 00:22:48.280 --rc geninfo_all_blocks=1 00:22:48.280 --rc geninfo_unexecuted_blocks=1 00:22:48.280 00:22:48.280 ' 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:22:48.280 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:48.280 --rc genhtml_branch_coverage=1 00:22:48.280 --rc genhtml_function_coverage=1 00:22:48.280 --rc genhtml_legend=1 00:22:48.280 --rc geninfo_all_blocks=1 00:22:48.280 --rc geninfo_unexecuted_blocks=1 00:22:48.280 00:22:48.280 ' 00:22:48.280 17:43:55 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # uname -s 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@15 -- # shopt -s extglob 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@5 -- # export PATH 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@51 -- # : 0 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:22:48.280 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@55 -- # have_pci_nics=0 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@11 -- # nvmftestinit 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@474 -- # prepare_net_devs 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@436 -- # local -g is_hw=no 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@438 -- # remove_spdk_ns 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@309 -- # xtrace_disable 00:22:48.280 17:43:56 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # pci_devs=() 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # net_devs=() 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # e810=() 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # local -ga e810 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # x722=() 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # local -ga x722 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # mlx=() 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # local -ga mlx 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:22:56.425 Found 0000:31:00.0 (0x8086 - 0x159b) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:22:56.425 Found 0000:31:00.1 (0x8086 - 0x159b) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:22:56.425 Found net devices under 0000:31:00.0: cvl_0_0 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:22:56.425 Found net devices under 0000:31:00.1: cvl_0_1 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # is_hw=yes 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:56.425 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:56.426 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:56.426 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.680 ms 00:22:56.426 00:22:56.426 --- 10.0.0.2 ping statistics --- 00:22:56.426 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:56.426 rtt min/avg/max/mdev = 0.680/0.680/0.680/0.000 ms 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:56.426 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:56.426 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.287 ms 00:22:56.426 00:22:56.426 --- 10.0.0.1 ping statistics --- 00:22:56.426 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:56.426 rtt min/avg/max/mdev = 0.287/0.287/0.287/0.000 ms 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@448 -- # return 0 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@12 -- # nvmfappstart -m 0xF 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@507 -- # nvmfpid=128659 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@508 -- # waitforlisten 128659 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@831 -- # '[' -z 128659 ']' 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:56.426 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:56.426 17:44:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.426 [2024-10-17 17:44:03.731357] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:22:56.426 [2024-10-17 17:44:03.731421] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:56.426 [2024-10-17 17:44:03.821704] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:56.426 [2024-10-17 17:44:03.874904] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:56.426 [2024-10-17 17:44:03.874957] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:56.426 [2024-10-17 17:44:03.874967] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:56.426 [2024-10-17 17:44:03.874976] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:56.426 [2024-10-17 17:44:03.874982] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:56.426 [2024-10-17 17:44:03.877108] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:56.426 [2024-10-17 17:44:03.877269] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:56.426 [2024-10-17 17:44:03.877416] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:56.426 [2024-10-17 17:44:03.877417] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:56.688 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:56.688 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@864 -- # return 0 00:22:56.688 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:22:56.688 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:56.688 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@14 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.949 [2024-10-17 17:44:04.613764] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@16 -- # rpc_cmd bdev_malloc_create 64 512 --name Malloc0 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.949 Malloc0 00:22:56.949 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@17 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 2 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@18 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@19 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.950 [2024-10-17 17:44:04.690885] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@21 -- # rpc_cmd nvmf_get_subsystems 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:56.950 [ 00:22:56.950 { 00:22:56.950 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:22:56.950 "subtype": "Discovery", 00:22:56.950 "listen_addresses": [], 00:22:56.950 "allow_any_host": true, 00:22:56.950 "hosts": [] 00:22:56.950 }, 00:22:56.950 { 00:22:56.950 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:22:56.950 "subtype": "NVMe", 00:22:56.950 "listen_addresses": [ 00:22:56.950 { 00:22:56.950 "trtype": "TCP", 00:22:56.950 "adrfam": "IPv4", 00:22:56.950 "traddr": "10.0.0.2", 00:22:56.950 "trsvcid": "4420" 00:22:56.950 } 00:22:56.950 ], 00:22:56.950 "allow_any_host": true, 00:22:56.950 "hosts": [], 00:22:56.950 "serial_number": "SPDK00000000000001", 00:22:56.950 "model_number": "SPDK bdev Controller", 00:22:56.950 "max_namespaces": 2, 00:22:56.950 "min_cntlid": 1, 00:22:56.950 "max_cntlid": 65519, 00:22:56.950 "namespaces": [ 00:22:56.950 { 00:22:56.950 "nsid": 1, 00:22:56.950 "bdev_name": "Malloc0", 00:22:56.950 "name": "Malloc0", 00:22:56.950 "nguid": "BF2CD23F106847A680B69F5606A89522", 00:22:56.950 "uuid": "bf2cd23f-1068-47a6-80b6-9f5606a89522" 00:22:56.950 } 00:22:56.950 ] 00:22:56.950 } 00:22:56.950 ] 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@23 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@24 -- # rm -f /tmp/aer_touch_file 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@33 -- # aerpid=128721 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@36 -- # waitforfile /tmp/aer_touch_file 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -n 2 -t /tmp/aer_touch_file 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1265 -- # local i=0 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 0 -lt 200 ']' 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1268 -- # i=1 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 1 -lt 200 ']' 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1268 -- # i=2 00:22:56.950 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:22:57.212 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:22:57.212 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 2 -lt 200 ']' 00:22:57.212 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1268 -- # i=3 00:22:57.212 17:44:04 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1276 -- # return 0 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@39 -- # rpc_cmd bdev_malloc_create 64 4096 --name Malloc1 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:57.212 Malloc1 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@40 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 2 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@41 -- # rpc_cmd nvmf_get_subsystems 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:57.212 Asynchronous Event Request test 00:22:57.212 Attaching to 10.0.0.2 00:22:57.212 Attached to 10.0.0.2 00:22:57.212 Registering asynchronous event callbacks... 00:22:57.212 Starting namespace attribute notice tests for all controllers... 00:22:57.212 10.0.0.2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:22:57.212 aer_cb - Changed Namespace 00:22:57.212 Cleaning up... 00:22:57.212 [ 00:22:57.212 { 00:22:57.212 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:22:57.212 "subtype": "Discovery", 00:22:57.212 "listen_addresses": [], 00:22:57.212 "allow_any_host": true, 00:22:57.212 "hosts": [] 00:22:57.212 }, 00:22:57.212 { 00:22:57.212 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:22:57.212 "subtype": "NVMe", 00:22:57.212 "listen_addresses": [ 00:22:57.212 { 00:22:57.212 "trtype": "TCP", 00:22:57.212 "adrfam": "IPv4", 00:22:57.212 "traddr": "10.0.0.2", 00:22:57.212 "trsvcid": "4420" 00:22:57.212 } 00:22:57.212 ], 00:22:57.212 "allow_any_host": true, 00:22:57.212 "hosts": [], 00:22:57.212 "serial_number": "SPDK00000000000001", 00:22:57.212 "model_number": "SPDK bdev Controller", 00:22:57.212 "max_namespaces": 2, 00:22:57.212 "min_cntlid": 1, 00:22:57.212 "max_cntlid": 65519, 00:22:57.212 "namespaces": [ 00:22:57.212 { 00:22:57.212 "nsid": 1, 00:22:57.212 "bdev_name": "Malloc0", 00:22:57.212 "name": "Malloc0", 00:22:57.212 "nguid": "BF2CD23F106847A680B69F5606A89522", 00:22:57.212 "uuid": "bf2cd23f-1068-47a6-80b6-9f5606a89522" 00:22:57.212 }, 00:22:57.212 { 00:22:57.212 "nsid": 2, 00:22:57.212 "bdev_name": "Malloc1", 00:22:57.212 "name": "Malloc1", 00:22:57.212 "nguid": "179D02953F054A6C87E75B23218FF544", 00:22:57.212 "uuid": "179d0295-3f05-4a6c-87e7-5b23218ff544" 00:22:57.212 } 00:22:57.212 ] 00:22:57.212 } 00:22:57.212 ] 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@43 -- # wait 128721 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@45 -- # rpc_cmd bdev_malloc_delete Malloc0 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.212 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@46 -- # rpc_cmd bdev_malloc_delete Malloc1 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@47 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@49 -- # trap - SIGINT SIGTERM EXIT 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@51 -- # nvmftestfini 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@121 -- # sync 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@124 -- # set +e 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:57.474 rmmod nvme_tcp 00:22:57.474 rmmod nvme_fabrics 00:22:57.474 rmmod nvme_keyring 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@128 -- # set -e 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@129 -- # return 0 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@515 -- # '[' -n 128659 ']' 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@516 -- # killprocess 128659 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@950 -- # '[' -z 128659 ']' 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@954 -- # kill -0 128659 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@955 -- # uname 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 128659 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@968 -- # echo 'killing process with pid 128659' 00:22:57.474 killing process with pid 128659 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@969 -- # kill 128659 00:22:57.474 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@974 -- # wait 128659 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@297 -- # iptr 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@789 -- # iptables-save 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@789 -- # iptables-restore 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:57.735 17:44:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:00.281 00:23:00.281 real 0m11.786s 00:23:00.281 user 0m8.608s 00:23:00.281 sys 0m6.383s 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:23:00.281 ************************************ 00:23:00.281 END TEST nvmf_aer 00:23:00.281 ************************************ 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@18 -- # run_test nvmf_async_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:23:00.281 ************************************ 00:23:00.281 START TEST nvmf_async_init 00:23:00.281 ************************************ 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:23:00.281 * Looking for test storage... 00:23:00.281 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1691 -- # lcov --version 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # IFS=.-: 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # read -ra ver1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # IFS=.-: 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # read -ra ver2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@338 -- # local 'op=<' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@340 -- # ver1_l=2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@341 -- # ver2_l=1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@344 -- # case "$op" in 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@345 -- # : 1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # decimal 1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # ver1[v]=1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # decimal 2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # ver2[v]=2 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # return 0 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:23:00.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:00.281 --rc genhtml_branch_coverage=1 00:23:00.281 --rc genhtml_function_coverage=1 00:23:00.281 --rc genhtml_legend=1 00:23:00.281 --rc geninfo_all_blocks=1 00:23:00.281 --rc geninfo_unexecuted_blocks=1 00:23:00.281 00:23:00.281 ' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:23:00.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:00.281 --rc genhtml_branch_coverage=1 00:23:00.281 --rc genhtml_function_coverage=1 00:23:00.281 --rc genhtml_legend=1 00:23:00.281 --rc geninfo_all_blocks=1 00:23:00.281 --rc geninfo_unexecuted_blocks=1 00:23:00.281 00:23:00.281 ' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:23:00.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:00.281 --rc genhtml_branch_coverage=1 00:23:00.281 --rc genhtml_function_coverage=1 00:23:00.281 --rc genhtml_legend=1 00:23:00.281 --rc geninfo_all_blocks=1 00:23:00.281 --rc geninfo_unexecuted_blocks=1 00:23:00.281 00:23:00.281 ' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:23:00.281 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:00.281 --rc genhtml_branch_coverage=1 00:23:00.281 --rc genhtml_function_coverage=1 00:23:00.281 --rc genhtml_legend=1 00:23:00.281 --rc geninfo_all_blocks=1 00:23:00.281 --rc geninfo_unexecuted_blocks=1 00:23:00.281 00:23:00.281 ' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # uname -s 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@15 -- # shopt -s extglob 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:00.281 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@5 -- # export PATH 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@51 -- # : 0 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:00.282 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@13 -- # null_bdev_size=1024 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@14 -- # null_block_size=512 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@15 -- # null_bdev=null0 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@16 -- # nvme_bdev=nvme0 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # uuidgen 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # tr -d - 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # nguid=a843e36cd58b4d3db8dd20d7ce0d7e63 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@22 -- # nvmftestinit 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@474 -- # prepare_net_devs 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@436 -- # local -g is_hw=no 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@438 -- # remove_spdk_ns 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@309 -- # xtrace_disable 00:23:00.282 17:44:07 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # pci_devs=() 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # net_devs=() 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # e810=() 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # local -ga e810 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # x722=() 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # local -ga x722 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # mlx=() 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # local -ga mlx 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:08.426 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:08.426 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:08.426 Found net devices under 0000:31:00.0: cvl_0_0 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:08.426 Found net devices under 0000:31:00.1: cvl_0_1 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # is_hw=yes 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:08.426 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:08.426 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.591 ms 00:23:08.426 00:23:08.426 --- 10.0.0.2 ping statistics --- 00:23:08.426 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:08.426 rtt min/avg/max/mdev = 0.591/0.591/0.591/0.000 ms 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:08.426 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:08.426 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.277 ms 00:23:08.426 00:23:08.426 --- 10.0.0.1 ping statistics --- 00:23:08.426 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:08.426 rtt min/avg/max/mdev = 0.277/0.277/0.277/0.000 ms 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:08.426 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@448 -- # return 0 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@23 -- # nvmfappstart -m 0x1 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@507 -- # nvmfpid=133081 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@508 -- # waitforlisten 133081 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@831 -- # '[' -z 133081 ']' 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:08.427 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:08.427 17:44:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.427 [2024-10-17 17:44:15.611272] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:23:08.427 [2024-10-17 17:44:15.611334] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:08.427 [2024-10-17 17:44:15.701044] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:08.427 [2024-10-17 17:44:15.753042] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:08.427 [2024-10-17 17:44:15.753091] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:08.427 [2024-10-17 17:44:15.753099] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:08.427 [2024-10-17 17:44:15.753106] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:08.427 [2024-10-17 17:44:15.753112] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:08.427 [2024-10-17 17:44:15.753915] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@864 -- # return 0 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@26 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.688 [2024-10-17 17:44:16.477180] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@27 -- # rpc_cmd bdev_null_create null0 1024 512 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.688 null0 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@28 -- # rpc_cmd bdev_wait_for_examine 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@29 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 -g a843e36cd58b4d3db8dd20d7ce0d7e63 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@31 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.688 [2024-10-17 17:44:16.537546] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@37 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode0 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.688 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.950 nvme0n1 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@41 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.950 [ 00:23:08.950 { 00:23:08.950 "name": "nvme0n1", 00:23:08.950 "aliases": [ 00:23:08.950 "a843e36c-d58b-4d3d-b8dd-20d7ce0d7e63" 00:23:08.950 ], 00:23:08.950 "product_name": "NVMe disk", 00:23:08.950 "block_size": 512, 00:23:08.950 "num_blocks": 2097152, 00:23:08.950 "uuid": "a843e36c-d58b-4d3d-b8dd-20d7ce0d7e63", 00:23:08.950 "numa_id": 0, 00:23:08.950 "assigned_rate_limits": { 00:23:08.950 "rw_ios_per_sec": 0, 00:23:08.950 "rw_mbytes_per_sec": 0, 00:23:08.950 "r_mbytes_per_sec": 0, 00:23:08.950 "w_mbytes_per_sec": 0 00:23:08.950 }, 00:23:08.950 "claimed": false, 00:23:08.950 "zoned": false, 00:23:08.950 "supported_io_types": { 00:23:08.950 "read": true, 00:23:08.950 "write": true, 00:23:08.950 "unmap": false, 00:23:08.950 "flush": true, 00:23:08.950 "reset": true, 00:23:08.950 "nvme_admin": true, 00:23:08.950 "nvme_io": true, 00:23:08.950 "nvme_io_md": false, 00:23:08.950 "write_zeroes": true, 00:23:08.950 "zcopy": false, 00:23:08.950 "get_zone_info": false, 00:23:08.950 "zone_management": false, 00:23:08.950 "zone_append": false, 00:23:08.950 "compare": true, 00:23:08.950 "compare_and_write": true, 00:23:08.950 "abort": true, 00:23:08.950 "seek_hole": false, 00:23:08.950 "seek_data": false, 00:23:08.950 "copy": true, 00:23:08.950 "nvme_iov_md": false 00:23:08.950 }, 00:23:08.950 "memory_domains": [ 00:23:08.950 { 00:23:08.950 "dma_device_id": "system", 00:23:08.950 "dma_device_type": 1 00:23:08.950 } 00:23:08.950 ], 00:23:08.950 "driver_specific": { 00:23:08.950 "nvme": [ 00:23:08.950 { 00:23:08.950 "trid": { 00:23:08.950 "trtype": "TCP", 00:23:08.950 "adrfam": "IPv4", 00:23:08.950 "traddr": "10.0.0.2", 00:23:08.950 "trsvcid": "4420", 00:23:08.950 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:23:08.950 }, 00:23:08.950 "ctrlr_data": { 00:23:08.950 "cntlid": 1, 00:23:08.950 "vendor_id": "0x8086", 00:23:08.950 "model_number": "SPDK bdev Controller", 00:23:08.950 "serial_number": "00000000000000000000", 00:23:08.950 "firmware_revision": "25.01", 00:23:08.950 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:23:08.950 "oacs": { 00:23:08.950 "security": 0, 00:23:08.950 "format": 0, 00:23:08.950 "firmware": 0, 00:23:08.950 "ns_manage": 0 00:23:08.950 }, 00:23:08.950 "multi_ctrlr": true, 00:23:08.950 "ana_reporting": false 00:23:08.950 }, 00:23:08.950 "vs": { 00:23:08.950 "nvme_version": "1.3" 00:23:08.950 }, 00:23:08.950 "ns_data": { 00:23:08.950 "id": 1, 00:23:08.950 "can_share": true 00:23:08.950 } 00:23:08.950 } 00:23:08.950 ], 00:23:08.950 "mp_policy": "active_passive" 00:23:08.950 } 00:23:08.950 } 00:23:08.950 ] 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@44 -- # rpc_cmd bdev_nvme_reset_controller nvme0 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.950 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:08.950 [2024-10-17 17:44:16.814023] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:08.950 [2024-10-17 17:44:16.814108] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147dc10 (9): Bad file descriptor 00:23:09.211 [2024-10-17 17:44:16.945860] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:23:09.211 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.211 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@47 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:23:09.211 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.211 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.211 [ 00:23:09.211 { 00:23:09.211 "name": "nvme0n1", 00:23:09.211 "aliases": [ 00:23:09.211 "a843e36c-d58b-4d3d-b8dd-20d7ce0d7e63" 00:23:09.211 ], 00:23:09.211 "product_name": "NVMe disk", 00:23:09.211 "block_size": 512, 00:23:09.211 "num_blocks": 2097152, 00:23:09.211 "uuid": "a843e36c-d58b-4d3d-b8dd-20d7ce0d7e63", 00:23:09.211 "numa_id": 0, 00:23:09.211 "assigned_rate_limits": { 00:23:09.211 "rw_ios_per_sec": 0, 00:23:09.211 "rw_mbytes_per_sec": 0, 00:23:09.211 "r_mbytes_per_sec": 0, 00:23:09.211 "w_mbytes_per_sec": 0 00:23:09.211 }, 00:23:09.211 "claimed": false, 00:23:09.212 "zoned": false, 00:23:09.212 "supported_io_types": { 00:23:09.212 "read": true, 00:23:09.212 "write": true, 00:23:09.212 "unmap": false, 00:23:09.212 "flush": true, 00:23:09.212 "reset": true, 00:23:09.212 "nvme_admin": true, 00:23:09.212 "nvme_io": true, 00:23:09.212 "nvme_io_md": false, 00:23:09.212 "write_zeroes": true, 00:23:09.212 "zcopy": false, 00:23:09.212 "get_zone_info": false, 00:23:09.212 "zone_management": false, 00:23:09.212 "zone_append": false, 00:23:09.212 "compare": true, 00:23:09.212 "compare_and_write": true, 00:23:09.212 "abort": true, 00:23:09.212 "seek_hole": false, 00:23:09.212 "seek_data": false, 00:23:09.212 "copy": true, 00:23:09.212 "nvme_iov_md": false 00:23:09.212 }, 00:23:09.212 "memory_domains": [ 00:23:09.212 { 00:23:09.212 "dma_device_id": "system", 00:23:09.212 "dma_device_type": 1 00:23:09.212 } 00:23:09.212 ], 00:23:09.212 "driver_specific": { 00:23:09.212 "nvme": [ 00:23:09.212 { 00:23:09.212 "trid": { 00:23:09.212 "trtype": "TCP", 00:23:09.212 "adrfam": "IPv4", 00:23:09.212 "traddr": "10.0.0.2", 00:23:09.212 "trsvcid": "4420", 00:23:09.212 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:23:09.212 }, 00:23:09.212 "ctrlr_data": { 00:23:09.212 "cntlid": 2, 00:23:09.212 "vendor_id": "0x8086", 00:23:09.212 "model_number": "SPDK bdev Controller", 00:23:09.212 "serial_number": "00000000000000000000", 00:23:09.212 "firmware_revision": "25.01", 00:23:09.212 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:23:09.212 "oacs": { 00:23:09.212 "security": 0, 00:23:09.212 "format": 0, 00:23:09.212 "firmware": 0, 00:23:09.212 "ns_manage": 0 00:23:09.212 }, 00:23:09.212 "multi_ctrlr": true, 00:23:09.212 "ana_reporting": false 00:23:09.212 }, 00:23:09.212 "vs": { 00:23:09.212 "nvme_version": "1.3" 00:23:09.212 }, 00:23:09.212 "ns_data": { 00:23:09.212 "id": 1, 00:23:09.212 "can_share": true 00:23:09.212 } 00:23:09.212 } 00:23:09.212 ], 00:23:09.212 "mp_policy": "active_passive" 00:23:09.212 } 00:23:09.212 } 00:23:09.212 ] 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@50 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # mktemp 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # key_path=/tmp/tmp.jLIXuCJlE4 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@54 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:23:09.212 17:44:16 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@55 -- # chmod 0600 /tmp/tmp.jLIXuCJlE4 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@56 -- # rpc_cmd keyring_file_add_key key0 /tmp/tmp.jLIXuCJlE4 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@57 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode0 --disable 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@58 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 --secure-channel 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.212 [2024-10-17 17:44:17.034701] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:23:09.212 [2024-10-17 17:44:17.034867] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@60 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host1 --psk key0 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@66 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4421 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.212 [2024-10-17 17:44:17.058772] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:09.212 nvme0n1 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@70 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.212 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.475 [ 00:23:09.475 { 00:23:09.475 "name": "nvme0n1", 00:23:09.475 "aliases": [ 00:23:09.476 "a843e36c-d58b-4d3d-b8dd-20d7ce0d7e63" 00:23:09.476 ], 00:23:09.476 "product_name": "NVMe disk", 00:23:09.476 "block_size": 512, 00:23:09.476 "num_blocks": 2097152, 00:23:09.476 "uuid": "a843e36c-d58b-4d3d-b8dd-20d7ce0d7e63", 00:23:09.476 "numa_id": 0, 00:23:09.476 "assigned_rate_limits": { 00:23:09.476 "rw_ios_per_sec": 0, 00:23:09.476 "rw_mbytes_per_sec": 0, 00:23:09.476 "r_mbytes_per_sec": 0, 00:23:09.476 "w_mbytes_per_sec": 0 00:23:09.476 }, 00:23:09.476 "claimed": false, 00:23:09.476 "zoned": false, 00:23:09.476 "supported_io_types": { 00:23:09.476 "read": true, 00:23:09.476 "write": true, 00:23:09.476 "unmap": false, 00:23:09.476 "flush": true, 00:23:09.476 "reset": true, 00:23:09.476 "nvme_admin": true, 00:23:09.476 "nvme_io": true, 00:23:09.476 "nvme_io_md": false, 00:23:09.476 "write_zeroes": true, 00:23:09.476 "zcopy": false, 00:23:09.476 "get_zone_info": false, 00:23:09.476 "zone_management": false, 00:23:09.476 "zone_append": false, 00:23:09.476 "compare": true, 00:23:09.476 "compare_and_write": true, 00:23:09.476 "abort": true, 00:23:09.476 "seek_hole": false, 00:23:09.476 "seek_data": false, 00:23:09.476 "copy": true, 00:23:09.476 "nvme_iov_md": false 00:23:09.476 }, 00:23:09.476 "memory_domains": [ 00:23:09.476 { 00:23:09.476 "dma_device_id": "system", 00:23:09.476 "dma_device_type": 1 00:23:09.476 } 00:23:09.476 ], 00:23:09.476 "driver_specific": { 00:23:09.476 "nvme": [ 00:23:09.476 { 00:23:09.476 "trid": { 00:23:09.476 "trtype": "TCP", 00:23:09.476 "adrfam": "IPv4", 00:23:09.476 "traddr": "10.0.0.2", 00:23:09.476 "trsvcid": "4421", 00:23:09.476 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:23:09.476 }, 00:23:09.476 "ctrlr_data": { 00:23:09.476 "cntlid": 3, 00:23:09.476 "vendor_id": "0x8086", 00:23:09.476 "model_number": "SPDK bdev Controller", 00:23:09.476 "serial_number": "00000000000000000000", 00:23:09.476 "firmware_revision": "25.01", 00:23:09.476 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:23:09.476 "oacs": { 00:23:09.476 "security": 0, 00:23:09.476 "format": 0, 00:23:09.476 "firmware": 0, 00:23:09.476 "ns_manage": 0 00:23:09.476 }, 00:23:09.476 "multi_ctrlr": true, 00:23:09.476 "ana_reporting": false 00:23:09.476 }, 00:23:09.476 "vs": { 00:23:09.476 "nvme_version": "1.3" 00:23:09.476 }, 00:23:09.476 "ns_data": { 00:23:09.476 "id": 1, 00:23:09.476 "can_share": true 00:23:09.476 } 00:23:09.476 } 00:23:09.476 ], 00:23:09.476 "mp_policy": "active_passive" 00:23:09.476 } 00:23:09.476 } 00:23:09.476 ] 00:23:09.476 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.476 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@73 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:09.476 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.476 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:09.476 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.476 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@76 -- # rm -f /tmp/tmp.jLIXuCJlE4 00:23:09.476 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@78 -- # trap - SIGINT SIGTERM EXIT 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@79 -- # nvmftestfini 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@514 -- # nvmfcleanup 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@121 -- # sync 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@124 -- # set +e 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:09.477 rmmod nvme_tcp 00:23:09.477 rmmod nvme_fabrics 00:23:09.477 rmmod nvme_keyring 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@128 -- # set -e 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@129 -- # return 0 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@515 -- # '[' -n 133081 ']' 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@516 -- # killprocess 133081 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@950 -- # '[' -z 133081 ']' 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@954 -- # kill -0 133081 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@955 -- # uname 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 133081 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 133081' 00:23:09.477 killing process with pid 133081 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@969 -- # kill 133081 00:23:09.477 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@974 -- # wait 133081 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@297 -- # iptr 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@789 -- # iptables-save 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@789 -- # iptables-restore 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:09.743 17:44:17 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:11.658 17:44:19 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:11.658 00:23:11.658 real 0m11.887s 00:23:11.658 user 0m4.249s 00:23:11.658 sys 0m6.190s 00:23:11.658 17:44:19 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:11.658 17:44:19 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:23:11.658 ************************************ 00:23:11.658 END TEST nvmf_async_init 00:23:11.658 ************************************ 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@19 -- # run_test dma /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:23:11.920 ************************************ 00:23:11.920 START TEST dma 00:23:11.920 ************************************ 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:23:11.920 * Looking for test storage... 00:23:11.920 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1691 -- # lcov --version 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # IFS=.-: 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # read -ra ver1 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # IFS=.-: 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # read -ra ver2 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@338 -- # local 'op=<' 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@340 -- # ver1_l=2 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@341 -- # ver2_l=1 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@344 -- # case "$op" in 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@345 -- # : 1 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # decimal 1 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=1 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 1 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # ver1[v]=1 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # decimal 2 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=2 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:11.920 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 2 00:23:12.181 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # ver2[v]=2 00:23:12.181 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:12.181 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:12.181 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # return 0 00:23:12.181 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:23:12.182 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.182 --rc genhtml_branch_coverage=1 00:23:12.182 --rc genhtml_function_coverage=1 00:23:12.182 --rc genhtml_legend=1 00:23:12.182 --rc geninfo_all_blocks=1 00:23:12.182 --rc geninfo_unexecuted_blocks=1 00:23:12.182 00:23:12.182 ' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:23:12.182 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.182 --rc genhtml_branch_coverage=1 00:23:12.182 --rc genhtml_function_coverage=1 00:23:12.182 --rc genhtml_legend=1 00:23:12.182 --rc geninfo_all_blocks=1 00:23:12.182 --rc geninfo_unexecuted_blocks=1 00:23:12.182 00:23:12.182 ' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:23:12.182 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.182 --rc genhtml_branch_coverage=1 00:23:12.182 --rc genhtml_function_coverage=1 00:23:12.182 --rc genhtml_legend=1 00:23:12.182 --rc geninfo_all_blocks=1 00:23:12.182 --rc geninfo_unexecuted_blocks=1 00:23:12.182 00:23:12.182 ' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:23:12.182 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.182 --rc genhtml_branch_coverage=1 00:23:12.182 --rc genhtml_function_coverage=1 00:23:12.182 --rc genhtml_legend=1 00:23:12.182 --rc geninfo_all_blocks=1 00:23:12.182 --rc geninfo_unexecuted_blocks=1 00:23:12.182 00:23:12.182 ' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- host/dma.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # uname -s 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@15 -- # shopt -s extglob 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- paths/export.sh@5 -- # export PATH 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@51 -- # : 0 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:12.182 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- host/dma.sh@12 -- # '[' tcp '!=' rdma ']' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- host/dma.sh@13 -- # exit 0 00:23:12.182 00:23:12.182 real 0m0.242s 00:23:12.182 user 0m0.135s 00:23:12.182 sys 0m0.123s 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@10 -- # set +x 00:23:12.182 ************************************ 00:23:12.182 END TEST dma 00:23:12.182 ************************************ 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@22 -- # run_test nvmf_identify /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:23:12.182 ************************************ 00:23:12.182 START TEST nvmf_identify 00:23:12.182 ************************************ 00:23:12.182 17:44:19 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:23:12.182 * Looking for test storage... 00:23:12.182 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:12.182 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:23:12.182 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1691 -- # lcov --version 00:23:12.182 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # IFS=.-: 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # read -ra ver1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # IFS=.-: 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # read -ra ver2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@338 -- # local 'op=<' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@340 -- # ver1_l=2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@341 -- # ver2_l=1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@344 -- # case "$op" in 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@345 -- # : 1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # decimal 1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # ver1[v]=1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # decimal 2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # ver2[v]=2 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # return 0 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:23:12.444 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.444 --rc genhtml_branch_coverage=1 00:23:12.444 --rc genhtml_function_coverage=1 00:23:12.444 --rc genhtml_legend=1 00:23:12.444 --rc geninfo_all_blocks=1 00:23:12.444 --rc geninfo_unexecuted_blocks=1 00:23:12.444 00:23:12.444 ' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:23:12.444 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.444 --rc genhtml_branch_coverage=1 00:23:12.444 --rc genhtml_function_coverage=1 00:23:12.444 --rc genhtml_legend=1 00:23:12.444 --rc geninfo_all_blocks=1 00:23:12.444 --rc geninfo_unexecuted_blocks=1 00:23:12.444 00:23:12.444 ' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:23:12.444 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.444 --rc genhtml_branch_coverage=1 00:23:12.444 --rc genhtml_function_coverage=1 00:23:12.444 --rc genhtml_legend=1 00:23:12.444 --rc geninfo_all_blocks=1 00:23:12.444 --rc geninfo_unexecuted_blocks=1 00:23:12.444 00:23:12.444 ' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:23:12.444 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:12.444 --rc genhtml_branch_coverage=1 00:23:12.444 --rc genhtml_function_coverage=1 00:23:12.444 --rc genhtml_legend=1 00:23:12.444 --rc geninfo_all_blocks=1 00:23:12.444 --rc geninfo_unexecuted_blocks=1 00:23:12.444 00:23:12.444 ' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # uname -s 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:12.444 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@15 -- # shopt -s extglob 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@5 -- # export PATH 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@51 -- # : 0 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:12.445 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@11 -- # MALLOC_BDEV_SIZE=64 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@14 -- # nvmftestinit 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@474 -- # prepare_net_devs 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@436 -- # local -g is_hw=no 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@438 -- # remove_spdk_ns 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@309 -- # xtrace_disable 00:23:12.445 17:44:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # pci_devs=() 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # net_devs=() 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # e810=() 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # local -ga e810 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # x722=() 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # local -ga x722 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # mlx=() 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # local -ga mlx 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:20.591 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:20.591 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:20.591 Found net devices under 0000:31:00.0: cvl_0_0 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:20.591 Found net devices under 0000:31:00.1: cvl_0_1 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # is_hw=yes 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:20.591 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:20.592 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:20.592 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.561 ms 00:23:20.592 00:23:20.592 --- 10.0.0.2 ping statistics --- 00:23:20.592 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:20.592 rtt min/avg/max/mdev = 0.561/0.561/0.561/0.000 ms 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:20.592 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:20.592 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.276 ms 00:23:20.592 00:23:20.592 --- 10.0.0.1 ping statistics --- 00:23:20.592 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:20.592 rtt min/avg/max/mdev = 0.276/0.276/0.276/0.000 ms 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@448 -- # return 0 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@16 -- # timing_enter start_nvmf_tgt 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@19 -- # nvmfpid=137829 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@21 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@18 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@23 -- # waitforlisten 137829 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@831 -- # '[' -z 137829 ']' 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:20.592 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:20.592 17:44:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:20.592 [2024-10-17 17:44:27.938640] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:23:20.592 [2024-10-17 17:44:27.938719] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:20.592 [2024-10-17 17:44:28.028442] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:23:20.592 [2024-10-17 17:44:28.083089] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:20.592 [2024-10-17 17:44:28.083140] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:20.592 [2024-10-17 17:44:28.083149] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:20.592 [2024-10-17 17:44:28.083156] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:20.592 [2024-10-17 17:44:28.083163] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:20.592 [2024-10-17 17:44:28.085641] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:20.592 [2024-10-17 17:44:28.085799] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:20.592 [2024-10-17 17:44:28.086091] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:23:20.592 [2024-10-17 17:44:28.086093] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:20.854 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:20.854 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@864 -- # return 0 00:23:20.854 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@24 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:23:20.854 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:20.854 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.116 [2024-10-17 17:44:28.777963] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@25 -- # timing_exit start_nvmf_tgt 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@27 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.116 Malloc0 00:23:21.116 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 --nguid ABCDEF0123456789ABCDEF0123456789 --eui64 ABCDEF0123456789 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.117 [2024-10-17 17:44:28.899782] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@35 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@37 -- # rpc_cmd nvmf_get_subsystems 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.117 [ 00:23:21.117 { 00:23:21.117 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:23:21.117 "subtype": "Discovery", 00:23:21.117 "listen_addresses": [ 00:23:21.117 { 00:23:21.117 "trtype": "TCP", 00:23:21.117 "adrfam": "IPv4", 00:23:21.117 "traddr": "10.0.0.2", 00:23:21.117 "trsvcid": "4420" 00:23:21.117 } 00:23:21.117 ], 00:23:21.117 "allow_any_host": true, 00:23:21.117 "hosts": [] 00:23:21.117 }, 00:23:21.117 { 00:23:21.117 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:21.117 "subtype": "NVMe", 00:23:21.117 "listen_addresses": [ 00:23:21.117 { 00:23:21.117 "trtype": "TCP", 00:23:21.117 "adrfam": "IPv4", 00:23:21.117 "traddr": "10.0.0.2", 00:23:21.117 "trsvcid": "4420" 00:23:21.117 } 00:23:21.117 ], 00:23:21.117 "allow_any_host": true, 00:23:21.117 "hosts": [], 00:23:21.117 "serial_number": "SPDK00000000000001", 00:23:21.117 "model_number": "SPDK bdev Controller", 00:23:21.117 "max_namespaces": 32, 00:23:21.117 "min_cntlid": 1, 00:23:21.117 "max_cntlid": 65519, 00:23:21.117 "namespaces": [ 00:23:21.117 { 00:23:21.117 "nsid": 1, 00:23:21.117 "bdev_name": "Malloc0", 00:23:21.117 "name": "Malloc0", 00:23:21.117 "nguid": "ABCDEF0123456789ABCDEF0123456789", 00:23:21.117 "eui64": "ABCDEF0123456789", 00:23:21.117 "uuid": "ea3bc356-631e-45e6-a23c-5dbdc018a299" 00:23:21.117 } 00:23:21.117 ] 00:23:21.117 } 00:23:21.117 ] 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.117 17:44:28 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' -L all 00:23:21.117 [2024-10-17 17:44:28.965220] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:23:21.117 [2024-10-17 17:44:28.965266] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid138181 ] 00:23:21.117 [2024-10-17 17:44:29.003935] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to connect adminq (no timeout) 00:23:21.117 [2024-10-17 17:44:29.004003] nvme_tcp.c:2349:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:23:21.117 [2024-10-17 17:44:29.004008] nvme_tcp.c:2353:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:23:21.117 [2024-10-17 17:44:29.004026] nvme_tcp.c:2374:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:23:21.117 [2024-10-17 17:44:29.004038] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:23:21.117 [2024-10-17 17:44:29.004749] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for connect adminq (no timeout) 00:23:21.117 [2024-10-17 17:44:29.004795] nvme_tcp.c:1566:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0xb2f620 0 00:23:21.117 [2024-10-17 17:44:29.010713] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:23:21.117 [2024-10-17 17:44:29.010730] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:23:21.117 [2024-10-17 17:44:29.010735] nvme_tcp.c:1612:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:23:21.117 [2024-10-17 17:44:29.010739] nvme_tcp.c:1613:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:23:21.117 [2024-10-17 17:44:29.010776] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.010783] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.010787] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.117 [2024-10-17 17:44:29.010805] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:23:21.117 [2024-10-17 17:44:29.010829] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.117 [2024-10-17 17:44:29.018706] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.117 [2024-10-17 17:44:29.018717] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.117 [2024-10-17 17:44:29.018721] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.018726] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.117 [2024-10-17 17:44:29.018740] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:23:21.117 [2024-10-17 17:44:29.018749] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read vs (no timeout) 00:23:21.117 [2024-10-17 17:44:29.018754] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read vs wait for vs (no timeout) 00:23:21.117 [2024-10-17 17:44:29.018771] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.018775] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.018779] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.117 [2024-10-17 17:44:29.018788] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.117 [2024-10-17 17:44:29.018803] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.117 [2024-10-17 17:44:29.019002] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.117 [2024-10-17 17:44:29.019009] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.117 [2024-10-17 17:44:29.019018] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.019022] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.117 [2024-10-17 17:44:29.019028] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read cap (no timeout) 00:23:21.117 [2024-10-17 17:44:29.019035] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read cap wait for cap (no timeout) 00:23:21.117 [2024-10-17 17:44:29.019043] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.019046] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.019050] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.117 [2024-10-17 17:44:29.019057] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.117 [2024-10-17 17:44:29.019067] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.117 [2024-10-17 17:44:29.019273] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.117 [2024-10-17 17:44:29.019280] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.117 [2024-10-17 17:44:29.019283] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.019287] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.117 [2024-10-17 17:44:29.019293] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to check en (no timeout) 00:23:21.117 [2024-10-17 17:44:29.019301] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to check en wait for cc (timeout 15000 ms) 00:23:21.117 [2024-10-17 17:44:29.019309] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.019312] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.019316] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.117 [2024-10-17 17:44:29.019323] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.117 [2024-10-17 17:44:29.019333] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.117 [2024-10-17 17:44:29.019543] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.117 [2024-10-17 17:44:29.019549] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.117 [2024-10-17 17:44:29.019552] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.117 [2024-10-17 17:44:29.019556] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.118 [2024-10-17 17:44:29.019562] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:23:21.118 [2024-10-17 17:44:29.019571] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.019575] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.019579] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.118 [2024-10-17 17:44:29.019586] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.118 [2024-10-17 17:44:29.019596] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.118 [2024-10-17 17:44:29.019798] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.118 [2024-10-17 17:44:29.019805] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.118 [2024-10-17 17:44:29.019808] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.019812] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.118 [2024-10-17 17:44:29.019817] nvme_ctrlr.c:3893:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CC.EN = 0 && CSTS.RDY = 0 00:23:21.118 [2024-10-17 17:44:29.019825] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to controller is disabled (timeout 15000 ms) 00:23:21.118 [2024-10-17 17:44:29.019833] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:23:21.118 [2024-10-17 17:44:29.019939] nvme_ctrlr.c:4091:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Setting CC.EN = 1 00:23:21.118 [2024-10-17 17:44:29.019944] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:23:21.118 [2024-10-17 17:44:29.019953] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.019957] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.019961] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.118 [2024-10-17 17:44:29.019967] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.118 [2024-10-17 17:44:29.019979] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.118 [2024-10-17 17:44:29.020193] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.118 [2024-10-17 17:44:29.020199] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.118 [2024-10-17 17:44:29.020203] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020207] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.118 [2024-10-17 17:44:29.020211] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:23:21.118 [2024-10-17 17:44:29.020221] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020224] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020228] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.118 [2024-10-17 17:44:29.020235] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.118 [2024-10-17 17:44:29.020245] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.118 [2024-10-17 17:44:29.020464] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.118 [2024-10-17 17:44:29.020470] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.118 [2024-10-17 17:44:29.020473] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020477] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.118 [2024-10-17 17:44:29.020482] nvme_ctrlr.c:3928:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:23:21.118 [2024-10-17 17:44:29.020486] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to reset admin queue (timeout 30000 ms) 00:23:21.118 [2024-10-17 17:44:29.020495] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to identify controller (no timeout) 00:23:21.118 [2024-10-17 17:44:29.020503] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for identify controller (timeout 30000 ms) 00:23:21.118 [2024-10-17 17:44:29.020513] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020517] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.118 [2024-10-17 17:44:29.020524] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.118 [2024-10-17 17:44:29.020537] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.118 [2024-10-17 17:44:29.020804] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.118 [2024-10-17 17:44:29.020811] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.118 [2024-10-17 17:44:29.020815] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020819] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0xb2f620): datao=0, datal=4096, cccid=0 00:23:21.118 [2024-10-17 17:44:29.020824] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0xb8f480) on tqpair(0xb2f620): expected_datao=0, payload_size=4096 00:23:21.118 [2024-10-17 17:44:29.020829] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020841] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.118 [2024-10-17 17:44:29.020846] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.066706] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.381 [2024-10-17 17:44:29.066738] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.381 [2024-10-17 17:44:29.066743] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.066747] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.381 [2024-10-17 17:44:29.066759] nvme_ctrlr.c:2077:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] transport max_xfer_size 4294967295 00:23:21.381 [2024-10-17 17:44:29.066764] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] MDTS max_xfer_size 131072 00:23:21.381 [2024-10-17 17:44:29.066769] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CNTLID 0x0001 00:23:21.381 [2024-10-17 17:44:29.066774] nvme_ctrlr.c:2108:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] transport max_sges 16 00:23:21.381 [2024-10-17 17:44:29.066779] nvme_ctrlr.c:2123:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] fuses compare and write: 1 00:23:21.381 [2024-10-17 17:44:29.066784] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to configure AER (timeout 30000 ms) 00:23:21.381 [2024-10-17 17:44:29.066794] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for configure aer (timeout 30000 ms) 00:23:21.381 [2024-10-17 17:44:29.066802] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.066806] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.066810] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.381 [2024-10-17 17:44:29.066820] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:23:21.381 [2024-10-17 17:44:29.066839] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.381 [2024-10-17 17:44:29.067073] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.381 [2024-10-17 17:44:29.067080] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.381 [2024-10-17 17:44:29.067083] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067087] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.381 [2024-10-17 17:44:29.067096] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067100] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067103] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0xb2f620) 00:23:21.381 [2024-10-17 17:44:29.067110] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.381 [2024-10-17 17:44:29.067117] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067121] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067130] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0xb2f620) 00:23:21.381 [2024-10-17 17:44:29.067136] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.381 [2024-10-17 17:44:29.067143] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067146] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067150] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0xb2f620) 00:23:21.381 [2024-10-17 17:44:29.067156] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.381 [2024-10-17 17:44:29.067162] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067166] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067169] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0xb2f620) 00:23:21.381 [2024-10-17 17:44:29.067175] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.381 [2024-10-17 17:44:29.067180] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to set keep alive timeout (timeout 30000 ms) 00:23:21.381 [2024-10-17 17:44:29.067193] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:23:21.381 [2024-10-17 17:44:29.067200] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.381 [2024-10-17 17:44:29.067204] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0xb2f620) 00:23:21.381 [2024-10-17 17:44:29.067211] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.381 [2024-10-17 17:44:29.067224] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f480, cid 0, qid 0 00:23:21.381 [2024-10-17 17:44:29.067229] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f600, cid 1, qid 0 00:23:21.381 [2024-10-17 17:44:29.067234] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f780, cid 2, qid 0 00:23:21.381 [2024-10-17 17:44:29.067239] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f900, cid 3, qid 0 00:23:21.381 [2024-10-17 17:44:29.067243] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8fa80, cid 4, qid 0 00:23:21.381 [2024-10-17 17:44:29.067503] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.382 [2024-10-17 17:44:29.067510] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.382 [2024-10-17 17:44:29.067513] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.067517] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8fa80) on tqpair=0xb2f620 00:23:21.382 [2024-10-17 17:44:29.067522] nvme_ctrlr.c:3046:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Sending keep alive every 5000000 us 00:23:21.382 [2024-10-17 17:44:29.067527] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to ready (no timeout) 00:23:21.382 [2024-10-17 17:44:29.067541] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.067545] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0xb2f620) 00:23:21.382 [2024-10-17 17:44:29.067551] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.382 [2024-10-17 17:44:29.067562] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8fa80, cid 4, qid 0 00:23:21.382 [2024-10-17 17:44:29.067743] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.382 [2024-10-17 17:44:29.067750] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.382 [2024-10-17 17:44:29.067756] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.067760] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0xb2f620): datao=0, datal=4096, cccid=4 00:23:21.382 [2024-10-17 17:44:29.067764] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0xb8fa80) on tqpair(0xb2f620): expected_datao=0, payload_size=4096 00:23:21.382 [2024-10-17 17:44:29.067769] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.067780] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.067785] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.067979] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.382 [2024-10-17 17:44:29.067985] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.382 [2024-10-17 17:44:29.067989] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.067992] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8fa80) on tqpair=0xb2f620 00:23:21.382 [2024-10-17 17:44:29.068006] nvme_ctrlr.c:4189:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Ctrlr already in ready state 00:23:21.382 [2024-10-17 17:44:29.068037] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068041] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0xb2f620) 00:23:21.382 [2024-10-17 17:44:29.068048] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.382 [2024-10-17 17:44:29.068055] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068059] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068062] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0xb2f620) 00:23:21.382 [2024-10-17 17:44:29.068069] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.382 [2024-10-17 17:44:29.068082] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8fa80, cid 4, qid 0 00:23:21.382 [2024-10-17 17:44:29.068087] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8fc00, cid 5, qid 0 00:23:21.382 [2024-10-17 17:44:29.068343] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.382 [2024-10-17 17:44:29.068349] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.382 [2024-10-17 17:44:29.068352] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068356] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0xb2f620): datao=0, datal=1024, cccid=4 00:23:21.382 [2024-10-17 17:44:29.068361] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0xb8fa80) on tqpair(0xb2f620): expected_datao=0, payload_size=1024 00:23:21.382 [2024-10-17 17:44:29.068365] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068372] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068375] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068381] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.382 [2024-10-17 17:44:29.068387] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.382 [2024-10-17 17:44:29.068390] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.068394] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8fc00) on tqpair=0xb2f620 00:23:21.382 [2024-10-17 17:44:29.114702] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.382 [2024-10-17 17:44:29.114716] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.382 [2024-10-17 17:44:29.114719] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.114724] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8fa80) on tqpair=0xb2f620 00:23:21.382 [2024-10-17 17:44:29.114746] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.114755] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0xb2f620) 00:23:21.382 [2024-10-17 17:44:29.114763] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:02ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.382 [2024-10-17 17:44:29.114785] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8fa80, cid 4, qid 0 00:23:21.382 [2024-10-17 17:44:29.115047] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.382 [2024-10-17 17:44:29.115053] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.382 [2024-10-17 17:44:29.115057] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.115061] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0xb2f620): datao=0, datal=3072, cccid=4 00:23:21.382 [2024-10-17 17:44:29.115065] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0xb8fa80) on tqpair(0xb2f620): expected_datao=0, payload_size=3072 00:23:21.382 [2024-10-17 17:44:29.115070] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.115084] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.115089] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.156889] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.382 [2024-10-17 17:44:29.156899] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.382 [2024-10-17 17:44:29.156902] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.156906] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8fa80) on tqpair=0xb2f620 00:23:21.382 [2024-10-17 17:44:29.156917] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.156920] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0xb2f620) 00:23:21.382 [2024-10-17 17:44:29.156928] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00010070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.382 [2024-10-17 17:44:29.156944] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8fa80, cid 4, qid 0 00:23:21.382 [2024-10-17 17:44:29.157158] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.382 [2024-10-17 17:44:29.157164] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.382 [2024-10-17 17:44:29.157168] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.157172] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0xb2f620): datao=0, datal=8, cccid=4 00:23:21.382 [2024-10-17 17:44:29.157176] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0xb8fa80) on tqpair(0xb2f620): expected_datao=0, payload_size=8 00:23:21.382 [2024-10-17 17:44:29.157181] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.157187] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.157191] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.197885] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.382 [2024-10-17 17:44:29.197895] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.382 [2024-10-17 17:44:29.197899] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.382 [2024-10-17 17:44:29.197903] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8fa80) on tqpair=0xb2f620 00:23:21.382 ===================================================== 00:23:21.382 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2014-08.org.nvmexpress.discovery 00:23:21.382 ===================================================== 00:23:21.382 Controller Capabilities/Features 00:23:21.382 ================================ 00:23:21.382 Vendor ID: 0000 00:23:21.382 Subsystem Vendor ID: 0000 00:23:21.382 Serial Number: .................... 00:23:21.382 Model Number: ........................................ 00:23:21.382 Firmware Version: 25.01 00:23:21.382 Recommended Arb Burst: 0 00:23:21.382 IEEE OUI Identifier: 00 00 00 00:23:21.382 Multi-path I/O 00:23:21.382 May have multiple subsystem ports: No 00:23:21.382 May have multiple controllers: No 00:23:21.382 Associated with SR-IOV VF: No 00:23:21.382 Max Data Transfer Size: 131072 00:23:21.382 Max Number of Namespaces: 0 00:23:21.382 Max Number of I/O Queues: 1024 00:23:21.382 NVMe Specification Version (VS): 1.3 00:23:21.382 NVMe Specification Version (Identify): 1.3 00:23:21.382 Maximum Queue Entries: 128 00:23:21.382 Contiguous Queues Required: Yes 00:23:21.382 Arbitration Mechanisms Supported 00:23:21.382 Weighted Round Robin: Not Supported 00:23:21.382 Vendor Specific: Not Supported 00:23:21.382 Reset Timeout: 15000 ms 00:23:21.382 Doorbell Stride: 4 bytes 00:23:21.382 NVM Subsystem Reset: Not Supported 00:23:21.382 Command Sets Supported 00:23:21.382 NVM Command Set: Supported 00:23:21.382 Boot Partition: Not Supported 00:23:21.382 Memory Page Size Minimum: 4096 bytes 00:23:21.382 Memory Page Size Maximum: 4096 bytes 00:23:21.382 Persistent Memory Region: Not Supported 00:23:21.382 Optional Asynchronous Events Supported 00:23:21.382 Namespace Attribute Notices: Not Supported 00:23:21.382 Firmware Activation Notices: Not Supported 00:23:21.382 ANA Change Notices: Not Supported 00:23:21.382 PLE Aggregate Log Change Notices: Not Supported 00:23:21.382 LBA Status Info Alert Notices: Not Supported 00:23:21.382 EGE Aggregate Log Change Notices: Not Supported 00:23:21.382 Normal NVM Subsystem Shutdown event: Not Supported 00:23:21.382 Zone Descriptor Change Notices: Not Supported 00:23:21.382 Discovery Log Change Notices: Supported 00:23:21.382 Controller Attributes 00:23:21.382 128-bit Host Identifier: Not Supported 00:23:21.382 Non-Operational Permissive Mode: Not Supported 00:23:21.382 NVM Sets: Not Supported 00:23:21.382 Read Recovery Levels: Not Supported 00:23:21.382 Endurance Groups: Not Supported 00:23:21.382 Predictable Latency Mode: Not Supported 00:23:21.382 Traffic Based Keep ALive: Not Supported 00:23:21.382 Namespace Granularity: Not Supported 00:23:21.382 SQ Associations: Not Supported 00:23:21.382 UUID List: Not Supported 00:23:21.382 Multi-Domain Subsystem: Not Supported 00:23:21.383 Fixed Capacity Management: Not Supported 00:23:21.383 Variable Capacity Management: Not Supported 00:23:21.383 Delete Endurance Group: Not Supported 00:23:21.383 Delete NVM Set: Not Supported 00:23:21.383 Extended LBA Formats Supported: Not Supported 00:23:21.383 Flexible Data Placement Supported: Not Supported 00:23:21.383 00:23:21.383 Controller Memory Buffer Support 00:23:21.383 ================================ 00:23:21.383 Supported: No 00:23:21.383 00:23:21.383 Persistent Memory Region Support 00:23:21.383 ================================ 00:23:21.383 Supported: No 00:23:21.383 00:23:21.383 Admin Command Set Attributes 00:23:21.383 ============================ 00:23:21.383 Security Send/Receive: Not Supported 00:23:21.383 Format NVM: Not Supported 00:23:21.383 Firmware Activate/Download: Not Supported 00:23:21.383 Namespace Management: Not Supported 00:23:21.383 Device Self-Test: Not Supported 00:23:21.383 Directives: Not Supported 00:23:21.383 NVMe-MI: Not Supported 00:23:21.383 Virtualization Management: Not Supported 00:23:21.383 Doorbell Buffer Config: Not Supported 00:23:21.383 Get LBA Status Capability: Not Supported 00:23:21.383 Command & Feature Lockdown Capability: Not Supported 00:23:21.383 Abort Command Limit: 1 00:23:21.383 Async Event Request Limit: 4 00:23:21.383 Number of Firmware Slots: N/A 00:23:21.383 Firmware Slot 1 Read-Only: N/A 00:23:21.383 Firmware Activation Without Reset: N/A 00:23:21.383 Multiple Update Detection Support: N/A 00:23:21.383 Firmware Update Granularity: No Information Provided 00:23:21.383 Per-Namespace SMART Log: No 00:23:21.383 Asymmetric Namespace Access Log Page: Not Supported 00:23:21.383 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:23:21.383 Command Effects Log Page: Not Supported 00:23:21.383 Get Log Page Extended Data: Supported 00:23:21.383 Telemetry Log Pages: Not Supported 00:23:21.383 Persistent Event Log Pages: Not Supported 00:23:21.383 Supported Log Pages Log Page: May Support 00:23:21.383 Commands Supported & Effects Log Page: Not Supported 00:23:21.383 Feature Identifiers & Effects Log Page:May Support 00:23:21.383 NVMe-MI Commands & Effects Log Page: May Support 00:23:21.383 Data Area 4 for Telemetry Log: Not Supported 00:23:21.383 Error Log Page Entries Supported: 128 00:23:21.383 Keep Alive: Not Supported 00:23:21.383 00:23:21.383 NVM Command Set Attributes 00:23:21.383 ========================== 00:23:21.383 Submission Queue Entry Size 00:23:21.383 Max: 1 00:23:21.383 Min: 1 00:23:21.383 Completion Queue Entry Size 00:23:21.383 Max: 1 00:23:21.383 Min: 1 00:23:21.383 Number of Namespaces: 0 00:23:21.383 Compare Command: Not Supported 00:23:21.383 Write Uncorrectable Command: Not Supported 00:23:21.383 Dataset Management Command: Not Supported 00:23:21.383 Write Zeroes Command: Not Supported 00:23:21.383 Set Features Save Field: Not Supported 00:23:21.383 Reservations: Not Supported 00:23:21.383 Timestamp: Not Supported 00:23:21.383 Copy: Not Supported 00:23:21.383 Volatile Write Cache: Not Present 00:23:21.383 Atomic Write Unit (Normal): 1 00:23:21.383 Atomic Write Unit (PFail): 1 00:23:21.383 Atomic Compare & Write Unit: 1 00:23:21.383 Fused Compare & Write: Supported 00:23:21.383 Scatter-Gather List 00:23:21.383 SGL Command Set: Supported 00:23:21.383 SGL Keyed: Supported 00:23:21.383 SGL Bit Bucket Descriptor: Not Supported 00:23:21.383 SGL Metadata Pointer: Not Supported 00:23:21.383 Oversized SGL: Not Supported 00:23:21.383 SGL Metadata Address: Not Supported 00:23:21.383 SGL Offset: Supported 00:23:21.383 Transport SGL Data Block: Not Supported 00:23:21.383 Replay Protected Memory Block: Not Supported 00:23:21.383 00:23:21.383 Firmware Slot Information 00:23:21.383 ========================= 00:23:21.383 Active slot: 0 00:23:21.383 00:23:21.383 00:23:21.383 Error Log 00:23:21.383 ========= 00:23:21.383 00:23:21.383 Active Namespaces 00:23:21.383 ================= 00:23:21.383 Discovery Log Page 00:23:21.383 ================== 00:23:21.383 Generation Counter: 2 00:23:21.383 Number of Records: 2 00:23:21.383 Record Format: 0 00:23:21.383 00:23:21.383 Discovery Log Entry 0 00:23:21.383 ---------------------- 00:23:21.383 Transport Type: 3 (TCP) 00:23:21.383 Address Family: 1 (IPv4) 00:23:21.383 Subsystem Type: 3 (Current Discovery Subsystem) 00:23:21.383 Entry Flags: 00:23:21.383 Duplicate Returned Information: 1 00:23:21.383 Explicit Persistent Connection Support for Discovery: 1 00:23:21.383 Transport Requirements: 00:23:21.383 Secure Channel: Not Required 00:23:21.383 Port ID: 0 (0x0000) 00:23:21.383 Controller ID: 65535 (0xffff) 00:23:21.383 Admin Max SQ Size: 128 00:23:21.383 Transport Service Identifier: 4420 00:23:21.383 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:23:21.383 Transport Address: 10.0.0.2 00:23:21.383 Discovery Log Entry 1 00:23:21.383 ---------------------- 00:23:21.383 Transport Type: 3 (TCP) 00:23:21.383 Address Family: 1 (IPv4) 00:23:21.383 Subsystem Type: 2 (NVM Subsystem) 00:23:21.383 Entry Flags: 00:23:21.383 Duplicate Returned Information: 0 00:23:21.383 Explicit Persistent Connection Support for Discovery: 0 00:23:21.383 Transport Requirements: 00:23:21.383 Secure Channel: Not Required 00:23:21.383 Port ID: 0 (0x0000) 00:23:21.383 Controller ID: 65535 (0xffff) 00:23:21.383 Admin Max SQ Size: 128 00:23:21.383 Transport Service Identifier: 4420 00:23:21.383 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:cnode1 00:23:21.383 Transport Address: 10.0.0.2 [2024-10-17 17:44:29.198011] nvme_ctrlr.c:4386:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Prepare to destruct SSD 00:23:21.383 [2024-10-17 17:44:29.198022] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f480) on tqpair=0xb2f620 00:23:21.383 [2024-10-17 17:44:29.198030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.383 [2024-10-17 17:44:29.198036] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f600) on tqpair=0xb2f620 00:23:21.383 [2024-10-17 17:44:29.198040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.383 [2024-10-17 17:44:29.198047] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f780) on tqpair=0xb2f620 00:23:21.383 [2024-10-17 17:44:29.198052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.383 [2024-10-17 17:44:29.198057] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f900) on tqpair=0xb2f620 00:23:21.383 [2024-10-17 17:44:29.198062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.383 [2024-10-17 17:44:29.198072] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198076] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198079] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0xb2f620) 00:23:21.383 [2024-10-17 17:44:29.198088] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.383 [2024-10-17 17:44:29.198104] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f900, cid 3, qid 0 00:23:21.383 [2024-10-17 17:44:29.198308] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.383 [2024-10-17 17:44:29.198314] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.383 [2024-10-17 17:44:29.198318] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198321] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f900) on tqpair=0xb2f620 00:23:21.383 [2024-10-17 17:44:29.198329] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198332] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198336] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0xb2f620) 00:23:21.383 [2024-10-17 17:44:29.198343] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.383 [2024-10-17 17:44:29.198355] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f900, cid 3, qid 0 00:23:21.383 [2024-10-17 17:44:29.198548] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.383 [2024-10-17 17:44:29.198554] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.383 [2024-10-17 17:44:29.198558] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198561] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f900) on tqpair=0xb2f620 00:23:21.383 [2024-10-17 17:44:29.198566] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] RTD3E = 0 us 00:23:21.383 [2024-10-17 17:44:29.198574] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] shutdown timeout = 10000 ms 00:23:21.383 [2024-10-17 17:44:29.198585] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198589] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.198592] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0xb2f620) 00:23:21.383 [2024-10-17 17:44:29.198599] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.383 [2024-10-17 17:44:29.198610] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0xb8f900, cid 3, qid 0 00:23:21.383 [2024-10-17 17:44:29.202701] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.383 [2024-10-17 17:44:29.202710] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.383 [2024-10-17 17:44:29.202714] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.383 [2024-10-17 17:44:29.202717] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0xb8f900) on tqpair=0xb2f620 00:23:21.383 [2024-10-17 17:44:29.202726] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] shutdown complete in 4 milliseconds 00:23:21.383 00:23:21.384 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -L all 00:23:21.384 [2024-10-17 17:44:29.250550] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:23:21.384 [2024-10-17 17:44:29.250625] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid138184 ] 00:23:21.384 [2024-10-17 17:44:29.283687] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to connect adminq (no timeout) 00:23:21.384 [2024-10-17 17:44:29.287753] nvme_tcp.c:2349:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:23:21.384 [2024-10-17 17:44:29.287758] nvme_tcp.c:2353:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:23:21.384 [2024-10-17 17:44:29.287774] nvme_tcp.c:2374:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:23:21.384 [2024-10-17 17:44:29.287785] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:23:21.384 [2024-10-17 17:44:29.288468] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for connect adminq (no timeout) 00:23:21.384 [2024-10-17 17:44:29.288504] nvme_tcp.c:1566:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x185a620 0 00:23:21.648 [2024-10-17 17:44:29.301702] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:23:21.648 [2024-10-17 17:44:29.301719] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:23:21.648 [2024-10-17 17:44:29.301724] nvme_tcp.c:1612:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:23:21.648 [2024-10-17 17:44:29.301727] nvme_tcp.c:1613:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:23:21.648 [2024-10-17 17:44:29.301759] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.301765] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.301769] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.301784] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:23:21.648 [2024-10-17 17:44:29.301809] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.309703] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.648 [2024-10-17 17:44:29.309713] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.648 [2024-10-17 17:44:29.309717] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.309721] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.648 [2024-10-17 17:44:29.309734] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:23:21.648 [2024-10-17 17:44:29.309742] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read vs (no timeout) 00:23:21.648 [2024-10-17 17:44:29.309748] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read vs wait for vs (no timeout) 00:23:21.648 [2024-10-17 17:44:29.309762] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.309767] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.309770] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.309779] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.648 [2024-10-17 17:44:29.309796] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.309993] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.648 [2024-10-17 17:44:29.310000] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.648 [2024-10-17 17:44:29.310003] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310008] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.648 [2024-10-17 17:44:29.310013] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read cap (no timeout) 00:23:21.648 [2024-10-17 17:44:29.310022] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read cap wait for cap (no timeout) 00:23:21.648 [2024-10-17 17:44:29.310029] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310033] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310036] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.310043] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.648 [2024-10-17 17:44:29.310054] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.310273] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.648 [2024-10-17 17:44:29.310279] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.648 [2024-10-17 17:44:29.310283] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310287] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.648 [2024-10-17 17:44:29.310292] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to check en (no timeout) 00:23:21.648 [2024-10-17 17:44:29.310301] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to check en wait for cc (timeout 15000 ms) 00:23:21.648 [2024-10-17 17:44:29.310308] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310312] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310315] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.310322] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.648 [2024-10-17 17:44:29.310333] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.310541] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.648 [2024-10-17 17:44:29.310547] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.648 [2024-10-17 17:44:29.310551] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310555] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.648 [2024-10-17 17:44:29.310560] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:23:21.648 [2024-10-17 17:44:29.310570] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310574] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310577] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.310584] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.648 [2024-10-17 17:44:29.310594] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.310778] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.648 [2024-10-17 17:44:29.310785] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.648 [2024-10-17 17:44:29.310791] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310795] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.648 [2024-10-17 17:44:29.310800] nvme_ctrlr.c:3893:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CC.EN = 0 && CSTS.RDY = 0 00:23:21.648 [2024-10-17 17:44:29.310805] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to controller is disabled (timeout 15000 ms) 00:23:21.648 [2024-10-17 17:44:29.310813] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:23:21.648 [2024-10-17 17:44:29.310918] nvme_ctrlr.c:4091:nvme_ctrlr_process_init: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Setting CC.EN = 1 00:23:21.648 [2024-10-17 17:44:29.310922] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:23:21.648 [2024-10-17 17:44:29.310931] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310935] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.310938] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.310945] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.648 [2024-10-17 17:44:29.310956] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.311162] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.648 [2024-10-17 17:44:29.311168] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.648 [2024-10-17 17:44:29.311172] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.311176] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.648 [2024-10-17 17:44:29.311180] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:23:21.648 [2024-10-17 17:44:29.311190] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.311194] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.311198] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.311204] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.648 [2024-10-17 17:44:29.311215] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.311430] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.648 [2024-10-17 17:44:29.311436] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.648 [2024-10-17 17:44:29.311440] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.311444] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.648 [2024-10-17 17:44:29.311448] nvme_ctrlr.c:3928:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:23:21.648 [2024-10-17 17:44:29.311453] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to reset admin queue (timeout 30000 ms) 00:23:21.648 [2024-10-17 17:44:29.311461] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify controller (no timeout) 00:23:21.648 [2024-10-17 17:44:29.311470] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify controller (timeout 30000 ms) 00:23:21.648 [2024-10-17 17:44:29.311479] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.311483] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.648 [2024-10-17 17:44:29.311490] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.648 [2024-10-17 17:44:29.311503] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.648 [2024-10-17 17:44:29.311738] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.648 [2024-10-17 17:44:29.311745] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.648 [2024-10-17 17:44:29.311749] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.648 [2024-10-17 17:44:29.311753] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=4096, cccid=0 00:23:21.648 [2024-10-17 17:44:29.311758] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18ba480) on tqpair(0x185a620): expected_datao=0, payload_size=4096 00:23:21.648 [2024-10-17 17:44:29.311762] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.311777] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.311782] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.352924] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.649 [2024-10-17 17:44:29.352938] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.649 [2024-10-17 17:44:29.352941] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.352946] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.649 [2024-10-17 17:44:29.352956] nvme_ctrlr.c:2077:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] transport max_xfer_size 4294967295 00:23:21.649 [2024-10-17 17:44:29.352961] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] MDTS max_xfer_size 131072 00:23:21.649 [2024-10-17 17:44:29.352966] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CNTLID 0x0001 00:23:21.649 [2024-10-17 17:44:29.352970] nvme_ctrlr.c:2108:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] transport max_sges 16 00:23:21.649 [2024-10-17 17:44:29.352975] nvme_ctrlr.c:2123:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] fuses compare and write: 1 00:23:21.649 [2024-10-17 17:44:29.352980] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to configure AER (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.352989] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for configure aer (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.352997] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353001] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353005] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.353014] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:23:21.649 [2024-10-17 17:44:29.353029] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.649 [2024-10-17 17:44:29.353172] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.649 [2024-10-17 17:44:29.353179] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.649 [2024-10-17 17:44:29.353183] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353187] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.649 [2024-10-17 17:44:29.353194] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353198] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353202] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.353208] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.649 [2024-10-17 17:44:29.353214] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353223] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353227] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.353233] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.649 [2024-10-17 17:44:29.353240] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353243] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353247] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.353253] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.649 [2024-10-17 17:44:29.353259] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353263] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353266] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.353272] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.649 [2024-10-17 17:44:29.353277] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set keep alive timeout (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.353290] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.353297] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353301] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.353308] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.649 [2024-10-17 17:44:29.353320] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba480, cid 0, qid 0 00:23:21.649 [2024-10-17 17:44:29.353326] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba600, cid 1, qid 0 00:23:21.649 [2024-10-17 17:44:29.353331] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba780, cid 2, qid 0 00:23:21.649 [2024-10-17 17:44:29.353335] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.649 [2024-10-17 17:44:29.353340] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baa80, cid 4, qid 0 00:23:21.649 [2024-10-17 17:44:29.353575] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.649 [2024-10-17 17:44:29.353581] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.649 [2024-10-17 17:44:29.353585] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353589] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baa80) on tqpair=0x185a620 00:23:21.649 [2024-10-17 17:44:29.353594] nvme_ctrlr.c:3046:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Sending keep alive every 5000000 us 00:23:21.649 [2024-10-17 17:44:29.353599] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify controller iocs specific (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.353608] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set number of queues (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.353617] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for set number of queues (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.353623] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353627] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353631] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.353640] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:4 cdw10:00000007 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:23:21.649 [2024-10-17 17:44:29.353650] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baa80, cid 4, qid 0 00:23:21.649 [2024-10-17 17:44:29.353908] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.649 [2024-10-17 17:44:29.353915] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.649 [2024-10-17 17:44:29.353918] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.353922] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baa80) on tqpair=0x185a620 00:23:21.649 [2024-10-17 17:44:29.353992] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify active ns (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.354002] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify active ns (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.354010] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354013] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.354020] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.649 [2024-10-17 17:44:29.354031] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baa80, cid 4, qid 0 00:23:21.649 [2024-10-17 17:44:29.354252] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.649 [2024-10-17 17:44:29.354259] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.649 [2024-10-17 17:44:29.354263] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354266] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=4096, cccid=4 00:23:21.649 [2024-10-17 17:44:29.354271] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18baa80) on tqpair(0x185a620): expected_datao=0, payload_size=4096 00:23:21.649 [2024-10-17 17:44:29.354276] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354283] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354287] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354461] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.649 [2024-10-17 17:44:29.354468] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.649 [2024-10-17 17:44:29.354471] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354475] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baa80) on tqpair=0x185a620 00:23:21.649 [2024-10-17 17:44:29.354485] nvme_ctrlr.c:4722:spdk_nvme_ctrlr_get_ns: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Namespace 1 was added 00:23:21.649 [2024-10-17 17:44:29.354503] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify ns (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.354513] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify ns (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.354520] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354524] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x185a620) 00:23:21.649 [2024-10-17 17:44:29.354530] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000000 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.649 [2024-10-17 17:44:29.354541] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baa80, cid 4, qid 0 00:23:21.649 [2024-10-17 17:44:29.354744] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.649 [2024-10-17 17:44:29.354751] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.649 [2024-10-17 17:44:29.354754] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354761] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=4096, cccid=4 00:23:21.649 [2024-10-17 17:44:29.354766] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18baa80) on tqpair(0x185a620): expected_datao=0, payload_size=4096 00:23:21.649 [2024-10-17 17:44:29.354770] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354777] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354781] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354966] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.649 [2024-10-17 17:44:29.354972] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.649 [2024-10-17 17:44:29.354976] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.354980] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baa80) on tqpair=0x185a620 00:23:21.649 [2024-10-17 17:44:29.354993] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.355003] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:23:21.649 [2024-10-17 17:44:29.355010] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.649 [2024-10-17 17:44:29.355014] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.355021] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.355031] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baa80, cid 4, qid 0 00:23:21.650 [2024-10-17 17:44:29.355259] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.650 [2024-10-17 17:44:29.355265] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.650 [2024-10-17 17:44:29.355269] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355272] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=4096, cccid=4 00:23:21.650 [2024-10-17 17:44:29.355277] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18baa80) on tqpair(0x185a620): expected_datao=0, payload_size=4096 00:23:21.650 [2024-10-17 17:44:29.355281] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355288] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355291] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355472] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.650 [2024-10-17 17:44:29.355478] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.650 [2024-10-17 17:44:29.355482] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355486] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baa80) on tqpair=0x185a620 00:23:21.650 [2024-10-17 17:44:29.355493] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify ns iocs specific (timeout 30000 ms) 00:23:21.650 [2024-10-17 17:44:29.355502] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set supported log pages (timeout 30000 ms) 00:23:21.650 [2024-10-17 17:44:29.355511] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set supported features (timeout 30000 ms) 00:23:21.650 [2024-10-17 17:44:29.355517] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set host behavior support feature (timeout 30000 ms) 00:23:21.650 [2024-10-17 17:44:29.355522] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set doorbell buffer config (timeout 30000 ms) 00:23:21.650 [2024-10-17 17:44:29.355527] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set host ID (timeout 30000 ms) 00:23:21.650 [2024-10-17 17:44:29.355535] nvme_ctrlr.c:3134:nvme_ctrlr_set_host_id: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] NVMe-oF transport - not sending Set Features - Host ID 00:23:21.650 [2024-10-17 17:44:29.355540] nvme_ctrlr.c:1557:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to transport ready (timeout 30000 ms) 00:23:21.650 [2024-10-17 17:44:29.355545] nvme_ctrlr.c:1563:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to ready (no timeout) 00:23:21.650 [2024-10-17 17:44:29.355562] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355566] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.355573] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:4 cdw10:00000001 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.355579] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355583] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.355587] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.355593] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:23:21.650 [2024-10-17 17:44:29.355605] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baa80, cid 4, qid 0 00:23:21.650 [2024-10-17 17:44:29.355610] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18bac00, cid 5, qid 0 00:23:21.650 [2024-10-17 17:44:29.359701] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.650 [2024-10-17 17:44:29.359710] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.650 [2024-10-17 17:44:29.359713] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.359717] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baa80) on tqpair=0x185a620 00:23:21.650 [2024-10-17 17:44:29.359724] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.650 [2024-10-17 17:44:29.359730] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.650 [2024-10-17 17:44:29.359734] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.359738] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18bac00) on tqpair=0x185a620 00:23:21.650 [2024-10-17 17:44:29.359748] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.359752] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.359758] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:5 cdw10:00000002 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.359770] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18bac00, cid 5, qid 0 00:23:21.650 [2024-10-17 17:44:29.359958] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.650 [2024-10-17 17:44:29.359965] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.650 [2024-10-17 17:44:29.359968] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.359972] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18bac00) on tqpair=0x185a620 00:23:21.650 [2024-10-17 17:44:29.359982] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.359986] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.359992] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:5 cdw10:00000004 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.360002] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18bac00, cid 5, qid 0 00:23:21.650 [2024-10-17 17:44:29.360196] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.650 [2024-10-17 17:44:29.360202] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.650 [2024-10-17 17:44:29.360209] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360213] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18bac00) on tqpair=0x185a620 00:23:21.650 [2024-10-17 17:44:29.360222] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360226] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.360232] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:5 cdw10:00000007 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.360243] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18bac00, cid 5, qid 0 00:23:21.650 [2024-10-17 17:44:29.360460] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.650 [2024-10-17 17:44:29.360466] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.650 [2024-10-17 17:44:29.360469] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360473] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18bac00) on tqpair=0x185a620 00:23:21.650 [2024-10-17 17:44:29.360489] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360494] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.360500] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:5 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.360508] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360512] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.360518] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:ffffffff cdw10:007f0002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.360525] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360529] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=6 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.360535] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:6 nsid:ffffffff cdw10:007f0003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.360545] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360549] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x185a620) 00:23:21.650 [2024-10-17 17:44:29.360555] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:7 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.650 [2024-10-17 17:44:29.360567] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18bac00, cid 5, qid 0 00:23:21.650 [2024-10-17 17:44:29.360572] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baa80, cid 4, qid 0 00:23:21.650 [2024-10-17 17:44:29.360577] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18bad80, cid 6, qid 0 00:23:21.650 [2024-10-17 17:44:29.360582] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baf00, cid 7, qid 0 00:23:21.650 [2024-10-17 17:44:29.360867] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.650 [2024-10-17 17:44:29.360875] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.650 [2024-10-17 17:44:29.360878] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.360882] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=8192, cccid=5 00:23:21.650 [2024-10-17 17:44:29.360887] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18bac00) on tqpair(0x185a620): expected_datao=0, payload_size=8192 00:23:21.650 [2024-10-17 17:44:29.360891] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361002] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361009] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361015] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.650 [2024-10-17 17:44:29.361021] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.650 [2024-10-17 17:44:29.361024] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361028] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=512, cccid=4 00:23:21.650 [2024-10-17 17:44:29.361033] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18baa80) on tqpair(0x185a620): expected_datao=0, payload_size=512 00:23:21.650 [2024-10-17 17:44:29.361037] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361043] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361047] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361053] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.650 [2024-10-17 17:44:29.361059] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.650 [2024-10-17 17:44:29.361062] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361066] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=512, cccid=6 00:23:21.650 [2024-10-17 17:44:29.361070] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18bad80) on tqpair(0x185a620): expected_datao=0, payload_size=512 00:23:21.650 [2024-10-17 17:44:29.361074] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361081] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361085] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.650 [2024-10-17 17:44:29.361090] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:23:21.650 [2024-10-17 17:44:29.361096] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:23:21.651 [2024-10-17 17:44:29.361100] nvme_tcp.c:1730:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361103] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x185a620): datao=0, datal=4096, cccid=7 00:23:21.651 [2024-10-17 17:44:29.361108] nvme_tcp.c:1742:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x18baf00) on tqpair(0x185a620): expected_datao=0, payload_size=4096 00:23:21.651 [2024-10-17 17:44:29.361112] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361130] nvme_tcp.c:1532:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361134] nvme_tcp.c:1323:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361295] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.651 [2024-10-17 17:44:29.361302] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.651 [2024-10-17 17:44:29.361305] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361309] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18bac00) on tqpair=0x185a620 00:23:21.651 [2024-10-17 17:44:29.361322] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.651 [2024-10-17 17:44:29.361329] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.651 [2024-10-17 17:44:29.361332] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361336] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baa80) on tqpair=0x185a620 00:23:21.651 [2024-10-17 17:44:29.361348] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.651 [2024-10-17 17:44:29.361354] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.651 [2024-10-17 17:44:29.361357] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361361] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18bad80) on tqpair=0x185a620 00:23:21.651 [2024-10-17 17:44:29.361368] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.651 [2024-10-17 17:44:29.361374] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.651 [2024-10-17 17:44:29.361379] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.651 [2024-10-17 17:44:29.361383] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baf00) on tqpair=0x185a620 00:23:21.651 ===================================================== 00:23:21.651 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:21.651 ===================================================== 00:23:21.651 Controller Capabilities/Features 00:23:21.651 ================================ 00:23:21.651 Vendor ID: 8086 00:23:21.651 Subsystem Vendor ID: 8086 00:23:21.651 Serial Number: SPDK00000000000001 00:23:21.651 Model Number: SPDK bdev Controller 00:23:21.651 Firmware Version: 25.01 00:23:21.651 Recommended Arb Burst: 6 00:23:21.651 IEEE OUI Identifier: e4 d2 5c 00:23:21.651 Multi-path I/O 00:23:21.651 May have multiple subsystem ports: Yes 00:23:21.651 May have multiple controllers: Yes 00:23:21.651 Associated with SR-IOV VF: No 00:23:21.651 Max Data Transfer Size: 131072 00:23:21.651 Max Number of Namespaces: 32 00:23:21.651 Max Number of I/O Queues: 127 00:23:21.651 NVMe Specification Version (VS): 1.3 00:23:21.651 NVMe Specification Version (Identify): 1.3 00:23:21.651 Maximum Queue Entries: 128 00:23:21.651 Contiguous Queues Required: Yes 00:23:21.651 Arbitration Mechanisms Supported 00:23:21.651 Weighted Round Robin: Not Supported 00:23:21.651 Vendor Specific: Not Supported 00:23:21.651 Reset Timeout: 15000 ms 00:23:21.651 Doorbell Stride: 4 bytes 00:23:21.651 NVM Subsystem Reset: Not Supported 00:23:21.651 Command Sets Supported 00:23:21.651 NVM Command Set: Supported 00:23:21.651 Boot Partition: Not Supported 00:23:21.651 Memory Page Size Minimum: 4096 bytes 00:23:21.651 Memory Page Size Maximum: 4096 bytes 00:23:21.651 Persistent Memory Region: Not Supported 00:23:21.651 Optional Asynchronous Events Supported 00:23:21.651 Namespace Attribute Notices: Supported 00:23:21.651 Firmware Activation Notices: Not Supported 00:23:21.651 ANA Change Notices: Not Supported 00:23:21.651 PLE Aggregate Log Change Notices: Not Supported 00:23:21.651 LBA Status Info Alert Notices: Not Supported 00:23:21.651 EGE Aggregate Log Change Notices: Not Supported 00:23:21.651 Normal NVM Subsystem Shutdown event: Not Supported 00:23:21.651 Zone Descriptor Change Notices: Not Supported 00:23:21.651 Discovery Log Change Notices: Not Supported 00:23:21.651 Controller Attributes 00:23:21.651 128-bit Host Identifier: Supported 00:23:21.651 Non-Operational Permissive Mode: Not Supported 00:23:21.651 NVM Sets: Not Supported 00:23:21.651 Read Recovery Levels: Not Supported 00:23:21.651 Endurance Groups: Not Supported 00:23:21.651 Predictable Latency Mode: Not Supported 00:23:21.651 Traffic Based Keep ALive: Not Supported 00:23:21.651 Namespace Granularity: Not Supported 00:23:21.651 SQ Associations: Not Supported 00:23:21.651 UUID List: Not Supported 00:23:21.651 Multi-Domain Subsystem: Not Supported 00:23:21.651 Fixed Capacity Management: Not Supported 00:23:21.651 Variable Capacity Management: Not Supported 00:23:21.651 Delete Endurance Group: Not Supported 00:23:21.651 Delete NVM Set: Not Supported 00:23:21.651 Extended LBA Formats Supported: Not Supported 00:23:21.651 Flexible Data Placement Supported: Not Supported 00:23:21.651 00:23:21.651 Controller Memory Buffer Support 00:23:21.651 ================================ 00:23:21.651 Supported: No 00:23:21.651 00:23:21.651 Persistent Memory Region Support 00:23:21.651 ================================ 00:23:21.651 Supported: No 00:23:21.651 00:23:21.651 Admin Command Set Attributes 00:23:21.651 ============================ 00:23:21.651 Security Send/Receive: Not Supported 00:23:21.651 Format NVM: Not Supported 00:23:21.651 Firmware Activate/Download: Not Supported 00:23:21.651 Namespace Management: Not Supported 00:23:21.651 Device Self-Test: Not Supported 00:23:21.651 Directives: Not Supported 00:23:21.651 NVMe-MI: Not Supported 00:23:21.651 Virtualization Management: Not Supported 00:23:21.651 Doorbell Buffer Config: Not Supported 00:23:21.651 Get LBA Status Capability: Not Supported 00:23:21.651 Command & Feature Lockdown Capability: Not Supported 00:23:21.651 Abort Command Limit: 4 00:23:21.651 Async Event Request Limit: 4 00:23:21.651 Number of Firmware Slots: N/A 00:23:21.651 Firmware Slot 1 Read-Only: N/A 00:23:21.651 Firmware Activation Without Reset: N/A 00:23:21.651 Multiple Update Detection Support: N/A 00:23:21.651 Firmware Update Granularity: No Information Provided 00:23:21.651 Per-Namespace SMART Log: No 00:23:21.651 Asymmetric Namespace Access Log Page: Not Supported 00:23:21.651 Subsystem NQN: nqn.2016-06.io.spdk:cnode1 00:23:21.651 Command Effects Log Page: Supported 00:23:21.651 Get Log Page Extended Data: Supported 00:23:21.651 Telemetry Log Pages: Not Supported 00:23:21.651 Persistent Event Log Pages: Not Supported 00:23:21.651 Supported Log Pages Log Page: May Support 00:23:21.651 Commands Supported & Effects Log Page: Not Supported 00:23:21.651 Feature Identifiers & Effects Log Page:May Support 00:23:21.651 NVMe-MI Commands & Effects Log Page: May Support 00:23:21.651 Data Area 4 for Telemetry Log: Not Supported 00:23:21.651 Error Log Page Entries Supported: 128 00:23:21.651 Keep Alive: Supported 00:23:21.651 Keep Alive Granularity: 10000 ms 00:23:21.651 00:23:21.651 NVM Command Set Attributes 00:23:21.651 ========================== 00:23:21.651 Submission Queue Entry Size 00:23:21.651 Max: 64 00:23:21.651 Min: 64 00:23:21.651 Completion Queue Entry Size 00:23:21.651 Max: 16 00:23:21.651 Min: 16 00:23:21.651 Number of Namespaces: 32 00:23:21.651 Compare Command: Supported 00:23:21.651 Write Uncorrectable Command: Not Supported 00:23:21.651 Dataset Management Command: Supported 00:23:21.651 Write Zeroes Command: Supported 00:23:21.651 Set Features Save Field: Not Supported 00:23:21.651 Reservations: Supported 00:23:21.651 Timestamp: Not Supported 00:23:21.651 Copy: Supported 00:23:21.651 Volatile Write Cache: Present 00:23:21.651 Atomic Write Unit (Normal): 1 00:23:21.651 Atomic Write Unit (PFail): 1 00:23:21.651 Atomic Compare & Write Unit: 1 00:23:21.651 Fused Compare & Write: Supported 00:23:21.651 Scatter-Gather List 00:23:21.651 SGL Command Set: Supported 00:23:21.651 SGL Keyed: Supported 00:23:21.651 SGL Bit Bucket Descriptor: Not Supported 00:23:21.651 SGL Metadata Pointer: Not Supported 00:23:21.651 Oversized SGL: Not Supported 00:23:21.651 SGL Metadata Address: Not Supported 00:23:21.651 SGL Offset: Supported 00:23:21.651 Transport SGL Data Block: Not Supported 00:23:21.651 Replay Protected Memory Block: Not Supported 00:23:21.651 00:23:21.651 Firmware Slot Information 00:23:21.651 ========================= 00:23:21.651 Active slot: 1 00:23:21.651 Slot 1 Firmware Revision: 25.01 00:23:21.651 00:23:21.651 00:23:21.651 Commands Supported and Effects 00:23:21.651 ============================== 00:23:21.651 Admin Commands 00:23:21.651 -------------- 00:23:21.651 Get Log Page (02h): Supported 00:23:21.651 Identify (06h): Supported 00:23:21.651 Abort (08h): Supported 00:23:21.651 Set Features (09h): Supported 00:23:21.651 Get Features (0Ah): Supported 00:23:21.651 Asynchronous Event Request (0Ch): Supported 00:23:21.651 Keep Alive (18h): Supported 00:23:21.651 I/O Commands 00:23:21.651 ------------ 00:23:21.651 Flush (00h): Supported LBA-Change 00:23:21.651 Write (01h): Supported LBA-Change 00:23:21.651 Read (02h): Supported 00:23:21.651 Compare (05h): Supported 00:23:21.651 Write Zeroes (08h): Supported LBA-Change 00:23:21.651 Dataset Management (09h): Supported LBA-Change 00:23:21.651 Copy (19h): Supported LBA-Change 00:23:21.651 00:23:21.651 Error Log 00:23:21.652 ========= 00:23:21.652 00:23:21.652 Arbitration 00:23:21.652 =========== 00:23:21.652 Arbitration Burst: 1 00:23:21.652 00:23:21.652 Power Management 00:23:21.652 ================ 00:23:21.652 Number of Power States: 1 00:23:21.652 Current Power State: Power State #0 00:23:21.652 Power State #0: 00:23:21.652 Max Power: 0.00 W 00:23:21.652 Non-Operational State: Operational 00:23:21.652 Entry Latency: Not Reported 00:23:21.652 Exit Latency: Not Reported 00:23:21.652 Relative Read Throughput: 0 00:23:21.652 Relative Read Latency: 0 00:23:21.652 Relative Write Throughput: 0 00:23:21.652 Relative Write Latency: 0 00:23:21.652 Idle Power: Not Reported 00:23:21.652 Active Power: Not Reported 00:23:21.652 Non-Operational Permissive Mode: Not Supported 00:23:21.652 00:23:21.652 Health Information 00:23:21.652 ================== 00:23:21.652 Critical Warnings: 00:23:21.652 Available Spare Space: OK 00:23:21.652 Temperature: OK 00:23:21.652 Device Reliability: OK 00:23:21.652 Read Only: No 00:23:21.652 Volatile Memory Backup: OK 00:23:21.652 Current Temperature: 0 Kelvin (-273 Celsius) 00:23:21.652 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:23:21.652 Available Spare: 0% 00:23:21.652 Available Spare Threshold: 0% 00:23:21.652 Life Percentage Used:[2024-10-17 17:44:29.361490] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.361495] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.361502] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:7 cdw10:00000005 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.361514] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18baf00, cid 7, qid 0 00:23:21.652 [2024-10-17 17:44:29.361785] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.652 [2024-10-17 17:44:29.361792] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.652 [2024-10-17 17:44:29.361796] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.361800] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18baf00) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.361836] nvme_ctrlr.c:4386:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Prepare to destruct SSD 00:23:21.652 [2024-10-17 17:44:29.361846] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba480) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.361853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.652 [2024-10-17 17:44:29.361859] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba600) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.361864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.652 [2024-10-17 17:44:29.361869] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba780) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.361874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.652 [2024-10-17 17:44:29.361879] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.361884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:21.652 [2024-10-17 17:44:29.361892] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.361896] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.361900] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.361907] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.361919] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.652 [2024-10-17 17:44:29.362187] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.652 [2024-10-17 17:44:29.362194] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.652 [2024-10-17 17:44:29.362197] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362201] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.362209] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362212] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362216] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.362223] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.362236] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.652 [2024-10-17 17:44:29.362489] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.652 [2024-10-17 17:44:29.362498] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.652 [2024-10-17 17:44:29.362502] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362506] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.362510] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] RTD3E = 0 us 00:23:21.652 [2024-10-17 17:44:29.362515] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] shutdown timeout = 10000 ms 00:23:21.652 [2024-10-17 17:44:29.362525] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362529] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362532] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.362539] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.362549] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.652 [2024-10-17 17:44:29.362753] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.652 [2024-10-17 17:44:29.362760] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.652 [2024-10-17 17:44:29.362763] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362767] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.362777] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362782] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.362785] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.362792] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.362802] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.652 [2024-10-17 17:44:29.363042] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.652 [2024-10-17 17:44:29.363048] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.652 [2024-10-17 17:44:29.363052] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363056] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.363066] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363070] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363073] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.363080] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.363090] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.652 [2024-10-17 17:44:29.363295] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.652 [2024-10-17 17:44:29.363301] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.652 [2024-10-17 17:44:29.363304] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363308] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.363318] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363322] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363326] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.363332] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.363342] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.652 [2024-10-17 17:44:29.363546] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.652 [2024-10-17 17:44:29.363552] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.652 [2024-10-17 17:44:29.363556] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363559] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.652 [2024-10-17 17:44:29.363569] nvme_tcp.c: 800:nvme_tcp_build_contig_request: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363573] nvme_tcp.c: 977:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:23:21.652 [2024-10-17 17:44:29.363577] nvme_tcp.c: 986:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x185a620) 00:23:21.652 [2024-10-17 17:44:29.363584] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:21.652 [2024-10-17 17:44:29.363594] nvme_tcp.c: 951:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x18ba900, cid 3, qid 0 00:23:21.652 [2024-10-17 17:44:29.367703] nvme_tcp.c:1198:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:23:21.653 [2024-10-17 17:44:29.367711] nvme_tcp.c:1986:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:23:21.653 [2024-10-17 17:44:29.367715] nvme_tcp.c:1659:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:23:21.653 [2024-10-17 17:44:29.367719] nvme_tcp.c:1079:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x18ba900) on tqpair=0x185a620 00:23:21.653 [2024-10-17 17:44:29.367727] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] shutdown complete in 5 milliseconds 00:23:21.653 0% 00:23:21.653 Data Units Read: 0 00:23:21.653 Data Units Written: 0 00:23:21.653 Host Read Commands: 0 00:23:21.653 Host Write Commands: 0 00:23:21.653 Controller Busy Time: 0 minutes 00:23:21.653 Power Cycles: 0 00:23:21.653 Power On Hours: 0 hours 00:23:21.653 Unsafe Shutdowns: 0 00:23:21.653 Unrecoverable Media Errors: 0 00:23:21.653 Lifetime Error Log Entries: 0 00:23:21.653 Warning Temperature Time: 0 minutes 00:23:21.653 Critical Temperature Time: 0 minutes 00:23:21.653 00:23:21.653 Number of Queues 00:23:21.653 ================ 00:23:21.653 Number of I/O Submission Queues: 127 00:23:21.653 Number of I/O Completion Queues: 127 00:23:21.653 00:23:21.653 Active Namespaces 00:23:21.653 ================= 00:23:21.653 Namespace ID:1 00:23:21.653 Error Recovery Timeout: Unlimited 00:23:21.653 Command Set Identifier: NVM (00h) 00:23:21.653 Deallocate: Supported 00:23:21.653 Deallocated/Unwritten Error: Not Supported 00:23:21.653 Deallocated Read Value: Unknown 00:23:21.653 Deallocate in Write Zeroes: Not Supported 00:23:21.653 Deallocated Guard Field: 0xFFFF 00:23:21.653 Flush: Supported 00:23:21.653 Reservation: Supported 00:23:21.653 Namespace Sharing Capabilities: Multiple Controllers 00:23:21.653 Size (in LBAs): 131072 (0GiB) 00:23:21.653 Capacity (in LBAs): 131072 (0GiB) 00:23:21.653 Utilization (in LBAs): 131072 (0GiB) 00:23:21.653 NGUID: ABCDEF0123456789ABCDEF0123456789 00:23:21.653 EUI64: ABCDEF0123456789 00:23:21.653 UUID: ea3bc356-631e-45e6-a23c-5dbdc018a299 00:23:21.653 Thin Provisioning: Not Supported 00:23:21.653 Per-NS Atomic Units: Yes 00:23:21.653 Atomic Boundary Size (Normal): 0 00:23:21.653 Atomic Boundary Size (PFail): 0 00:23:21.653 Atomic Boundary Offset: 0 00:23:21.653 Maximum Single Source Range Length: 65535 00:23:21.653 Maximum Copy Length: 65535 00:23:21.653 Maximum Source Range Count: 1 00:23:21.653 NGUID/EUI64 Never Reused: No 00:23:21.653 Namespace Write Protected: No 00:23:21.653 Number of LBA Formats: 1 00:23:21.653 Current LBA Format: LBA Format #00 00:23:21.653 LBA Format #00: Data Size: 512 Metadata Size: 0 00:23:21.653 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@51 -- # sync 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@52 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@54 -- # trap - SIGINT SIGTERM EXIT 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@56 -- # nvmftestfini 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@514 -- # nvmfcleanup 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@121 -- # sync 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@124 -- # set +e 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:21.653 rmmod nvme_tcp 00:23:21.653 rmmod nvme_fabrics 00:23:21.653 rmmod nvme_keyring 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@128 -- # set -e 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@129 -- # return 0 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@515 -- # '[' -n 137829 ']' 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@516 -- # killprocess 137829 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@950 -- # '[' -z 137829 ']' 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@954 -- # kill -0 137829 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@955 -- # uname 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 137829 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@968 -- # echo 'killing process with pid 137829' 00:23:21.653 killing process with pid 137829 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@969 -- # kill 137829 00:23:21.653 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@974 -- # wait 137829 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@297 -- # iptr 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@789 -- # iptables-save 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@789 -- # iptables-restore 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:21.914 17:44:29 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:24.469 00:23:24.469 real 0m11.844s 00:23:24.469 user 0m8.781s 00:23:24.469 sys 0m6.279s 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:23:24.469 ************************************ 00:23:24.469 END TEST nvmf_identify 00:23:24.469 ************************************ 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@23 -- # run_test nvmf_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:23:24.469 ************************************ 00:23:24.469 START TEST nvmf_perf 00:23:24.469 ************************************ 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:23:24.469 * Looking for test storage... 00:23:24.469 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1691 -- # lcov --version 00:23:24.469 17:44:31 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # IFS=.-: 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # read -ra ver1 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # IFS=.-: 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # read -ra ver2 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@338 -- # local 'op=<' 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@340 -- # ver1_l=2 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@341 -- # ver2_l=1 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:24.469 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@344 -- # case "$op" in 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@345 -- # : 1 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # decimal 1 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=1 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 1 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # ver1[v]=1 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # decimal 2 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=2 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 2 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # ver2[v]=2 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # return 0 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:23:24.470 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:24.470 --rc genhtml_branch_coverage=1 00:23:24.470 --rc genhtml_function_coverage=1 00:23:24.470 --rc genhtml_legend=1 00:23:24.470 --rc geninfo_all_blocks=1 00:23:24.470 --rc geninfo_unexecuted_blocks=1 00:23:24.470 00:23:24.470 ' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:23:24.470 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:24.470 --rc genhtml_branch_coverage=1 00:23:24.470 --rc genhtml_function_coverage=1 00:23:24.470 --rc genhtml_legend=1 00:23:24.470 --rc geninfo_all_blocks=1 00:23:24.470 --rc geninfo_unexecuted_blocks=1 00:23:24.470 00:23:24.470 ' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:23:24.470 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:24.470 --rc genhtml_branch_coverage=1 00:23:24.470 --rc genhtml_function_coverage=1 00:23:24.470 --rc genhtml_legend=1 00:23:24.470 --rc geninfo_all_blocks=1 00:23:24.470 --rc geninfo_unexecuted_blocks=1 00:23:24.470 00:23:24.470 ' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:23:24.470 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:24.470 --rc genhtml_branch_coverage=1 00:23:24.470 --rc genhtml_function_coverage=1 00:23:24.470 --rc genhtml_legend=1 00:23:24.470 --rc geninfo_all_blocks=1 00:23:24.470 --rc geninfo_unexecuted_blocks=1 00:23:24.470 00:23:24.470 ' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # uname -s 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@15 -- # shopt -s extglob 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@5 -- # export PATH 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@51 -- # : 0 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:24.470 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@12 -- # MALLOC_BDEV_SIZE=64 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@17 -- # nvmftestinit 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@474 -- # prepare_net_devs 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@436 -- # local -g is_hw=no 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@438 -- # remove_spdk_ns 00:23:24.470 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:24.471 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:24.471 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:24.471 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:23:24.471 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:23:24.471 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@309 -- # xtrace_disable 00:23:24.471 17:44:32 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # pci_devs=() 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # net_devs=() 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # e810=() 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # local -ga e810 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # x722=() 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # local -ga x722 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # mlx=() 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # local -ga mlx 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:32.706 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:32.706 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:32.706 Found net devices under 0000:31:00.0: cvl_0_0 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:32.706 Found net devices under 0000:31:00.1: cvl_0_1 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # is_hw=yes 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:32.706 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:32.706 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.682 ms 00:23:32.706 00:23:32.706 --- 10.0.0.2 ping statistics --- 00:23:32.706 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:32.706 rtt min/avg/max/mdev = 0.682/0.682/0.682/0.000 ms 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:32.706 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:32.706 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.322 ms 00:23:32.706 00:23:32.706 --- 10.0.0.1 ping statistics --- 00:23:32.706 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:32.706 rtt min/avg/max/mdev = 0.322/0.322/0.322/0.000 ms 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@448 -- # return 0 00:23:32.706 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@18 -- # nvmfappstart -m 0xF 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@507 -- # nvmfpid=142522 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@508 -- # waitforlisten 142522 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@831 -- # '[' -z 142522 ']' 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:32.707 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:32.707 17:44:39 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:23:32.707 [2024-10-17 17:44:39.818606] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:23:32.707 [2024-10-17 17:44:39.818668] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:32.707 [2024-10-17 17:44:39.910741] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:23:32.707 [2024-10-17 17:44:39.964508] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:32.707 [2024-10-17 17:44:39.964560] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:32.707 [2024-10-17 17:44:39.964568] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:32.707 [2024-10-17 17:44:39.964575] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:32.707 [2024-10-17 17:44:39.964582] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:32.707 [2024-10-17 17:44:39.966725] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:32.707 [2024-10-17 17:44:39.966849] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:32.707 [2024-10-17 17:44:39.967008] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:23:32.707 [2024-10-17 17:44:39.967009] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:32.967 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:32.967 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@864 -- # return 0 00:23:32.967 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:23:32.967 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:32.967 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:23:32.967 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:32.968 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:23:32.968 17:44:40 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:23:33.541 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_get_config bdev 00:23:33.541 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # jq -r '.[].params | select(.name=="Nvme0").traddr' 00:23:33.541 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # local_nvme_trid=0000:65:00.0 00:23:33.541 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:23:33.802 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # bdevs=' Malloc0' 00:23:33.802 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@33 -- # '[' -n 0000:65:00.0 ']' 00:23:33.802 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@34 -- # bdevs=' Malloc0 Nvme0n1' 00:23:33.802 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@37 -- # '[' tcp == rdma ']' 00:23:33.802 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:23:34.063 [2024-10-17 17:44:41.821716] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:34.063 17:44:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:23:34.325 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:23:34.325 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:23:34.586 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:23:34.586 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:23:34.586 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:23:34.846 [2024-10-17 17:44:42.629741] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:34.846 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:23:35.105 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@52 -- # '[' -n 0000:65:00.0 ']' 00:23:35.105 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@53 -- # perf_app -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:65:00.0' 00:23:35.105 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@21 -- # '[' 0 -eq 1 ']' 00:23:35.105 17:44:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:65:00.0' 00:23:36.486 Initializing NVMe Controllers 00:23:36.486 Attached to NVMe Controller at 0000:65:00.0 [144d:a80a] 00:23:36.487 Associating PCIE (0000:65:00.0) NSID 1 with lcore 0 00:23:36.487 Initialization complete. Launching workers. 00:23:36.487 ======================================================== 00:23:36.487 Latency(us) 00:23:36.487 Device Information : IOPS MiB/s Average min max 00:23:36.487 PCIE (0000:65:00.0) NSID 1 from core 0: 78444.61 306.42 407.36 13.32 5111.02 00:23:36.487 ======================================================== 00:23:36.487 Total : 78444.61 306.42 407.36 13.32 5111.02 00:23:36.487 00:23:36.487 17:44:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:23:37.870 Initializing NVMe Controllers 00:23:37.870 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:37.870 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:23:37.870 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:23:37.870 Initialization complete. Launching workers. 00:23:37.870 ======================================================== 00:23:37.870 Latency(us) 00:23:37.870 Device Information : IOPS MiB/s Average min max 00:23:37.870 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 95.00 0.37 10776.09 110.05 45755.16 00:23:37.870 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 51.00 0.20 19698.13 4993.79 48000.90 00:23:37.870 ======================================================== 00:23:37.870 Total : 146.00 0.57 13892.69 110.05 48000.90 00:23:37.870 00:23:37.870 17:44:45 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 4096 -w randrw -M 50 -t 1 -HI -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:23:38.808 Initializing NVMe Controllers 00:23:38.808 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:38.808 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:23:38.808 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:23:38.808 Initialization complete. Launching workers. 00:23:38.808 ======================================================== 00:23:38.808 Latency(us) 00:23:38.808 Device Information : IOPS MiB/s Average min max 00:23:38.808 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 12712.00 49.66 2517.56 401.83 8059.49 00:23:38.808 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 3682.00 14.38 8734.10 4703.77 18033.21 00:23:38.808 ======================================================== 00:23:38.808 Total : 16394.00 64.04 3913.76 401.83 18033.21 00:23:38.808 00:23:38.808 17:44:46 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ e810 == \e\8\1\0 ]] 00:23:38.808 17:44:46 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ tcp == \r\d\m\a ]] 00:23:38.808 17:44:46 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -O 16384 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:23:41.350 Initializing NVMe Controllers 00:23:41.350 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:41.350 Controller IO queue size 128, less than required. 00:23:41.350 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:41.350 Controller IO queue size 128, less than required. 00:23:41.350 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:41.350 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:23:41.350 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:23:41.350 Initialization complete. Launching workers. 00:23:41.350 ======================================================== 00:23:41.350 Latency(us) 00:23:41.350 Device Information : IOPS MiB/s Average min max 00:23:41.350 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1972.71 493.18 65789.61 41328.80 106730.18 00:23:41.350 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 608.79 152.20 216995.23 56106.93 340290.57 00:23:41.350 ======================================================== 00:23:41.350 Total : 2581.50 645.38 101448.28 41328.80 340290.57 00:23:41.350 00:23:41.350 17:44:48 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 36964 -O 4096 -w randrw -M 50 -t 5 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0xf -P 4 00:23:41.350 No valid NVMe controllers or AIO or URING devices found 00:23:41.350 Initializing NVMe Controllers 00:23:41.350 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:41.350 Controller IO queue size 128, less than required. 00:23:41.350 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:41.350 WARNING: IO size 36964 (-o) is not a multiple of nsid 1 sector size 512. Removing this ns from test 00:23:41.350 Controller IO queue size 128, less than required. 00:23:41.350 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:41.350 WARNING: IO size 36964 (-o) is not a multiple of nsid 2 sector size 512. Removing this ns from test 00:23:41.350 WARNING: Some requested NVMe devices were skipped 00:23:41.350 17:44:49 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' --transport-stat 00:23:43.894 Initializing NVMe Controllers 00:23:43.894 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:43.894 Controller IO queue size 128, less than required. 00:23:43.894 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:43.894 Controller IO queue size 128, less than required. 00:23:43.894 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:23:43.894 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:23:43.894 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:23:43.894 Initialization complete. Launching workers. 00:23:43.894 00:23:43.894 ==================== 00:23:43.894 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 statistics: 00:23:43.894 TCP transport: 00:23:43.894 polls: 37141 00:23:43.894 idle_polls: 22144 00:23:43.894 sock_completions: 14997 00:23:43.894 nvme_completions: 7675 00:23:43.894 submitted_requests: 11498 00:23:43.894 queued_requests: 1 00:23:43.894 00:23:43.894 ==================== 00:23:43.894 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 statistics: 00:23:43.894 TCP transport: 00:23:43.894 polls: 38002 00:23:43.894 idle_polls: 22516 00:23:43.894 sock_completions: 15486 00:23:43.894 nvme_completions: 7185 00:23:43.894 submitted_requests: 10726 00:23:43.894 queued_requests: 1 00:23:43.894 ======================================================== 00:23:43.894 Latency(us) 00:23:43.894 Device Information : IOPS MiB/s Average min max 00:23:43.894 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1916.18 479.05 68121.02 40790.75 114246.27 00:23:43.894 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 1793.83 448.46 72557.52 25158.22 116013.60 00:23:43.894 ======================================================== 00:23:43.894 Total : 3710.01 927.50 70266.11 25158.22 116013.60 00:23:43.894 00:23:43.894 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@66 -- # sync 00:23:43.894 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@67 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@69 -- # '[' 0 -eq 1 ']' 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@112 -- # trap - SIGINT SIGTERM EXIT 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@114 -- # nvmftestfini 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@514 -- # nvmfcleanup 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@121 -- # sync 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@124 -- # set +e 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@125 -- # for i in {1..20} 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:23:44.155 rmmod nvme_tcp 00:23:44.155 rmmod nvme_fabrics 00:23:44.155 rmmod nvme_keyring 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@128 -- # set -e 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@129 -- # return 0 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@515 -- # '[' -n 142522 ']' 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@516 -- # killprocess 142522 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@950 -- # '[' -z 142522 ']' 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@954 -- # kill -0 142522 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@955 -- # uname 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 142522 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@968 -- # echo 'killing process with pid 142522' 00:23:44.155 killing process with pid 142522 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@969 -- # kill 142522 00:23:44.155 17:44:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@974 -- # wait 142522 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@297 -- # iptr 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@789 -- # iptables-save 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@789 -- # iptables-restore 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:46.066 17:44:53 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:23:48.611 00:23:48.611 real 0m24.120s 00:23:48.611 user 0m57.539s 00:23:48.611 sys 0m8.663s 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:23:48.611 ************************************ 00:23:48.611 END TEST nvmf_perf 00:23:48.611 ************************************ 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@24 -- # run_test nvmf_fio_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:23:48.611 ************************************ 00:23:48.611 START TEST nvmf_fio_host 00:23:48.611 ************************************ 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:23:48.611 * Looking for test storage... 00:23:48.611 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1691 -- # lcov --version 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # IFS=.-: 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # read -ra ver1 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # IFS=.-: 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # read -ra ver2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@338 -- # local 'op=<' 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@340 -- # ver1_l=2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@341 -- # ver2_l=1 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@344 -- # case "$op" in 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@345 -- # : 1 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # decimal 1 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=1 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 1 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # ver1[v]=1 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # decimal 2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # ver2[v]=2 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # return 0 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:23:48.611 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:48.611 --rc genhtml_branch_coverage=1 00:23:48.611 --rc genhtml_function_coverage=1 00:23:48.611 --rc genhtml_legend=1 00:23:48.611 --rc geninfo_all_blocks=1 00:23:48.611 --rc geninfo_unexecuted_blocks=1 00:23:48.611 00:23:48.611 ' 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:23:48.611 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:48.611 --rc genhtml_branch_coverage=1 00:23:48.611 --rc genhtml_function_coverage=1 00:23:48.611 --rc genhtml_legend=1 00:23:48.611 --rc geninfo_all_blocks=1 00:23:48.611 --rc geninfo_unexecuted_blocks=1 00:23:48.611 00:23:48.611 ' 00:23:48.611 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:23:48.611 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:48.611 --rc genhtml_branch_coverage=1 00:23:48.611 --rc genhtml_function_coverage=1 00:23:48.611 --rc genhtml_legend=1 00:23:48.611 --rc geninfo_all_blocks=1 00:23:48.611 --rc geninfo_unexecuted_blocks=1 00:23:48.611 00:23:48.612 ' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:23:48.612 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:48.612 --rc genhtml_branch_coverage=1 00:23:48.612 --rc genhtml_function_coverage=1 00:23:48.612 --rc genhtml_legend=1 00:23:48.612 --rc geninfo_all_blocks=1 00:23:48.612 --rc geninfo_unexecuted_blocks=1 00:23:48.612 00:23:48.612 ' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # uname -s 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@51 -- # : 0 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:48.612 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@14 -- # nvmftestinit 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@474 -- # prepare_net_devs 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@436 -- # local -g is_hw=no 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@438 -- # remove_spdk_ns 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@309 -- # xtrace_disable 00:23:48.612 17:44:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # pci_devs=() 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # net_devs=() 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # e810=() 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # local -ga e810 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # x722=() 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # local -ga x722 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # mlx=() 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # local -ga mlx 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:23:56.754 Found 0000:31:00.0 (0x8086 - 0x159b) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:23:56.754 Found 0000:31:00.1 (0x8086 - 0x159b) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:23:56.754 Found net devices under 0000:31:00.0: cvl_0_0 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:23:56.754 Found net devices under 0000:31:00.1: cvl_0_1 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # is_hw=yes 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:56.754 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:56.755 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:56.755 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.615 ms 00:23:56.755 00:23:56.755 --- 10.0.0.2 ping statistics --- 00:23:56.755 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:56.755 rtt min/avg/max/mdev = 0.615/0.615/0.615/0.000 ms 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:56.755 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:56.755 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.314 ms 00:23:56.755 00:23:56.755 --- 10.0.0.1 ping statistics --- 00:23:56.755 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:56.755 rtt min/avg/max/mdev = 0.314/0.314/0.314/0.000 ms 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@448 -- # return 0 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@16 -- # [[ y != y ]] 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@21 -- # timing_enter start_nvmf_tgt 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@24 -- # nvmfpid=149464 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@26 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@23 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@28 -- # waitforlisten 149464 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@831 -- # '[' -z 149464 ']' 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:56.755 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:56.755 17:45:03 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:23:56.755 [2024-10-17 17:45:04.029590] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:23:56.755 [2024-10-17 17:45:04.029657] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:56.755 [2024-10-17 17:45:04.124249] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:23:56.755 [2024-10-17 17:45:04.178366] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:56.755 [2024-10-17 17:45:04.178417] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:56.755 [2024-10-17 17:45:04.178426] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:56.755 [2024-10-17 17:45:04.178433] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:56.755 [2024-10-17 17:45:04.178440] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:56.755 [2024-10-17 17:45:04.181022] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:56.755 [2024-10-17 17:45:04.181182] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:56.755 [2024-10-17 17:45:04.181342] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:23:56.755 [2024-10-17 17:45:04.181342] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:23:57.017 17:45:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:57.017 17:45:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@864 -- # return 0 00:23:57.017 17:45:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:23:57.278 [2024-10-17 17:45:05.026006] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:57.278 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@30 -- # timing_exit start_nvmf_tgt 00:23:57.278 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:57.278 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:23:57.278 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:23:57.539 Malloc1 00:23:57.539 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:23:57.799 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:23:58.061 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:23:58.061 [2024-10-17 17:45:05.881645] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:58.061 17:45:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@38 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@41 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # local sanitizers 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # shift 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local asan_lib= 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libasan 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:23:58.322 17:45:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:23:58.584 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:23:58.584 fio-3.35 00:23:58.584 Starting 1 thread 00:24:01.150 00:24:01.150 test: (groupid=0, jobs=1): err= 0: pid=150299: Thu Oct 17 17:45:08 2024 00:24:01.150 read: IOPS=11.2k, BW=43.7MiB/s (45.8MB/s)(87.6MiB/2004msec) 00:24:01.150 slat (usec): min=2, max=242, avg= 2.15, stdev= 2.32 00:24:01.150 clat (usec): min=3152, max=9302, avg=6293.26, stdev=1250.76 00:24:01.150 lat (usec): min=3187, max=9304, avg=6295.41, stdev=1250.76 00:24:01.150 clat percentiles (usec): 00:24:01.150 | 1.00th=[ 4359], 5.00th=[ 4621], 10.00th=[ 4817], 20.00th=[ 5014], 00:24:01.150 | 30.00th=[ 5145], 40.00th=[ 5407], 50.00th=[ 6587], 60.00th=[ 7046], 00:24:01.150 | 70.00th=[ 7308], 80.00th=[ 7570], 90.00th=[ 7832], 95.00th=[ 8094], 00:24:01.150 | 99.00th=[ 8455], 99.50th=[ 8586], 99.90th=[ 8848], 99.95th=[ 9110], 00:24:01.150 | 99.99th=[ 9110] 00:24:01.150 bw ( KiB/s): min=36408, max=56128, per=99.86%, avg=44700.00, stdev=8870.96, samples=4 00:24:01.150 iops : min= 9102, max=14032, avg=11175.00, stdev=2217.74, samples=4 00:24:01.150 write: IOPS=11.1k, BW=43.5MiB/s (45.6MB/s)(87.2MiB/2004msec); 0 zone resets 00:24:01.150 slat (usec): min=2, max=220, avg= 2.23, stdev= 1.63 00:24:01.150 clat (usec): min=2399, max=7612, avg=5093.72, stdev=995.76 00:24:01.150 lat (usec): min=2414, max=7614, avg=5095.94, stdev=995.78 00:24:01.150 clat percentiles (usec): 00:24:01.150 | 1.00th=[ 3556], 5.00th=[ 3785], 10.00th=[ 3884], 20.00th=[ 4047], 00:24:01.150 | 30.00th=[ 4228], 40.00th=[ 4359], 50.00th=[ 5276], 60.00th=[ 5669], 00:24:01.150 | 70.00th=[ 5932], 80.00th=[ 6063], 90.00th=[ 6325], 95.00th=[ 6521], 00:24:01.150 | 99.00th=[ 6783], 99.50th=[ 6915], 99.90th=[ 7177], 99.95th=[ 7242], 00:24:01.150 | 99.99th=[ 7504] 00:24:01.150 bw ( KiB/s): min=37392, max=55736, per=99.97%, avg=44562.00, stdev=8619.02, samples=4 00:24:01.150 iops : min= 9348, max=13934, avg=11140.50, stdev=2154.75, samples=4 00:24:01.150 lat (msec) : 4=8.38%, 10=91.62% 00:24:01.150 cpu : usr=71.94%, sys=26.81%, ctx=53, majf=0, minf=9 00:24:01.150 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.9% 00:24:01.150 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:24:01.150 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:24:01.150 issued rwts: total=22425,22333,0,0 short=0,0,0,0 dropped=0,0,0,0 00:24:01.150 latency : target=0, window=0, percentile=100.00%, depth=128 00:24:01.150 00:24:01.150 Run status group 0 (all jobs): 00:24:01.150 READ: bw=43.7MiB/s (45.8MB/s), 43.7MiB/s-43.7MiB/s (45.8MB/s-45.8MB/s), io=87.6MiB (91.9MB), run=2004-2004msec 00:24:01.150 WRITE: bw=43.5MiB/s (45.6MB/s), 43.5MiB/s-43.5MiB/s (45.6MB/s-45.6MB/s), io=87.2MiB (91.5MB), run=2004-2004msec 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@45 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # local sanitizers 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # shift 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local asan_lib= 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libasan 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:24:01.150 17:45:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:24:01.411 test: (g=0): rw=randrw, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=spdk, iodepth=128 00:24:01.411 fio-3.35 00:24:01.411 Starting 1 thread 00:24:03.951 00:24:03.951 test: (groupid=0, jobs=1): err= 0: pid=150827: Thu Oct 17 17:45:11 2024 00:24:03.951 read: IOPS=9735, BW=152MiB/s (160MB/s)(305MiB/2006msec) 00:24:03.951 slat (usec): min=3, max=113, avg= 3.63, stdev= 1.66 00:24:03.951 clat (usec): min=1379, max=16981, avg=7922.96, stdev=1976.65 00:24:03.951 lat (usec): min=1382, max=16984, avg=7926.59, stdev=1976.87 00:24:03.951 clat percentiles (usec): 00:24:03.951 | 1.00th=[ 4228], 5.00th=[ 4948], 10.00th=[ 5473], 20.00th=[ 6128], 00:24:03.951 | 30.00th=[ 6718], 40.00th=[ 7177], 50.00th=[ 7701], 60.00th=[ 8356], 00:24:03.951 | 70.00th=[ 8979], 80.00th=[ 9896], 90.00th=[10683], 95.00th=[11076], 00:24:03.951 | 99.00th=[12649], 99.50th=[13173], 99.90th=[14746], 99.95th=[15270], 00:24:03.951 | 99.99th=[15795] 00:24:03.951 bw ( KiB/s): min=73216, max=84023, per=49.93%, avg=77781.75, stdev=4517.67, samples=4 00:24:03.951 iops : min= 4576, max= 5251, avg=4861.25, stdev=282.15, samples=4 00:24:03.951 write: IOPS=5763, BW=90.0MiB/s (94.4MB/s)(159MiB/1764msec); 0 zone resets 00:24:03.951 slat (usec): min=39, max=449, avg=41.06, stdev= 9.00 00:24:03.951 clat (usec): min=1975, max=17158, avg=9037.83, stdev=1359.33 00:24:03.951 lat (usec): min=2015, max=17295, avg=9078.90, stdev=1362.24 00:24:03.951 clat percentiles (usec): 00:24:03.951 | 1.00th=[ 6390], 5.00th=[ 7111], 10.00th=[ 7504], 20.00th=[ 7963], 00:24:03.951 | 30.00th=[ 8291], 40.00th=[ 8586], 50.00th=[ 8979], 60.00th=[ 9372], 00:24:03.951 | 70.00th=[ 9634], 80.00th=[10159], 90.00th=[10683], 95.00th=[11207], 00:24:03.951 | 99.00th=[12256], 99.50th=[13304], 99.90th=[16581], 99.95th=[16909], 00:24:03.951 | 99.99th=[17171] 00:24:03.951 bw ( KiB/s): min=76960, max=87345, per=87.87%, avg=81028.25, stdev=4434.26, samples=4 00:24:03.951 iops : min= 4810, max= 5459, avg=5064.25, stdev=277.11, samples=4 00:24:03.951 lat (msec) : 2=0.07%, 4=0.47%, 10=79.47%, 20=19.99% 00:24:03.951 cpu : usr=86.43%, sys=12.67%, ctx=11, majf=0, minf=25 00:24:03.951 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.4%, 32=0.8%, >=64=98.5% 00:24:03.951 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:24:03.951 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:24:03.951 issued rwts: total=19530,10166,0,0 short=0,0,0,0 dropped=0,0,0,0 00:24:03.951 latency : target=0, window=0, percentile=100.00%, depth=128 00:24:03.951 00:24:03.951 Run status group 0 (all jobs): 00:24:03.951 READ: bw=152MiB/s (160MB/s), 152MiB/s-152MiB/s (160MB/s-160MB/s), io=305MiB (320MB), run=2006-2006msec 00:24:03.951 WRITE: bw=90.0MiB/s (94.4MB/s), 90.0MiB/s-90.0MiB/s (94.4MB/s-94.4MB/s), io=159MiB (167MB), run=1764-1764msec 00:24:03.951 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:24:03.951 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@49 -- # '[' 0 -eq 1 ']' 00:24:03.951 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:24:03.951 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@85 -- # rm -f ./local-test-0-verify.state 00:24:03.951 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@86 -- # nvmftestfini 00:24:03.951 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@514 -- # nvmfcleanup 00:24:03.951 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@121 -- # sync 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@124 -- # set +e 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:03.952 rmmod nvme_tcp 00:24:03.952 rmmod nvme_fabrics 00:24:03.952 rmmod nvme_keyring 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@128 -- # set -e 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@129 -- # return 0 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@515 -- # '[' -n 149464 ']' 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@516 -- # killprocess 149464 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@950 -- # '[' -z 149464 ']' 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@954 -- # kill -0 149464 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@955 -- # uname 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 149464 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@968 -- # echo 'killing process with pid 149464' 00:24:03.952 killing process with pid 149464 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@969 -- # kill 149464 00:24:03.952 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@974 -- # wait 149464 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@297 -- # iptr 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@789 -- # iptables-save 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@789 -- # iptables-restore 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:04.212 17:45:11 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:06.121 17:45:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:06.121 00:24:06.121 real 0m17.933s 00:24:06.121 user 0m58.822s 00:24:06.121 sys 0m7.915s 00:24:06.121 17:45:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:06.121 17:45:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:24:06.121 ************************************ 00:24:06.121 END TEST nvmf_fio_host 00:24:06.121 ************************************ 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@25 -- # run_test nvmf_failover /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:24:06.381 ************************************ 00:24:06.381 START TEST nvmf_failover 00:24:06.381 ************************************ 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:24:06.381 * Looking for test storage... 00:24:06.381 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1691 -- # lcov --version 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # IFS=.-: 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # read -ra ver1 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # IFS=.-: 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # read -ra ver2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@338 -- # local 'op=<' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@340 -- # ver1_l=2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@341 -- # ver2_l=1 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@344 -- # case "$op" in 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@345 -- # : 1 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # decimal 1 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=1 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 1 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # ver1[v]=1 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # decimal 2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # ver2[v]=2 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # return 0 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:24:06.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:06.381 --rc genhtml_branch_coverage=1 00:24:06.381 --rc genhtml_function_coverage=1 00:24:06.381 --rc genhtml_legend=1 00:24:06.381 --rc geninfo_all_blocks=1 00:24:06.381 --rc geninfo_unexecuted_blocks=1 00:24:06.381 00:24:06.381 ' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:24:06.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:06.381 --rc genhtml_branch_coverage=1 00:24:06.381 --rc genhtml_function_coverage=1 00:24:06.381 --rc genhtml_legend=1 00:24:06.381 --rc geninfo_all_blocks=1 00:24:06.381 --rc geninfo_unexecuted_blocks=1 00:24:06.381 00:24:06.381 ' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:24:06.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:06.381 --rc genhtml_branch_coverage=1 00:24:06.381 --rc genhtml_function_coverage=1 00:24:06.381 --rc genhtml_legend=1 00:24:06.381 --rc geninfo_all_blocks=1 00:24:06.381 --rc geninfo_unexecuted_blocks=1 00:24:06.381 00:24:06.381 ' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:24:06.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:06.381 --rc genhtml_branch_coverage=1 00:24:06.381 --rc genhtml_function_coverage=1 00:24:06.381 --rc genhtml_legend=1 00:24:06.381 --rc geninfo_all_blocks=1 00:24:06.381 --rc geninfo_unexecuted_blocks=1 00:24:06.381 00:24:06.381 ' 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:06.381 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # uname -s 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@15 -- # shopt -s extglob 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@5 -- # export PATH 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@51 -- # : 0 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:06.642 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@11 -- # MALLOC_BDEV_SIZE=64 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@18 -- # nvmftestinit 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@474 -- # prepare_net_devs 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@436 -- # local -g is_hw=no 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@438 -- # remove_spdk_ns 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@309 -- # xtrace_disable 00:24:06.642 17:45:14 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # pci_devs=() 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # net_devs=() 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # e810=() 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # local -ga e810 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # x722=() 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # local -ga x722 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # mlx=() 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # local -ga mlx 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:24:14.778 Found 0000:31:00.0 (0x8086 - 0x159b) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:24:14.778 Found 0000:31:00.1 (0x8086 - 0x159b) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:24:14.778 Found net devices under 0000:31:00.0: cvl_0_0 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:24:14.778 Found net devices under 0000:31:00.1: cvl_0_1 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # is_hw=yes 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:14.778 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:14.778 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.639 ms 00:24:14.778 00:24:14.778 --- 10.0.0.2 ping statistics --- 00:24:14.778 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:14.778 rtt min/avg/max/mdev = 0.639/0.639/0.639/0.000 ms 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:14.778 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:14.778 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.287 ms 00:24:14.778 00:24:14.778 --- 10.0.0.1 ping statistics --- 00:24:14.778 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:14.778 rtt min/avg/max/mdev = 0.287/0.287/0.287/0.000 ms 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@448 -- # return 0 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:14.778 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@20 -- # nvmfappstart -m 0xE 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@724 -- # xtrace_disable 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@507 -- # nvmfpid=156043 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@508 -- # waitforlisten 156043 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@831 -- # '[' -z 156043 ']' 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:14.779 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:14.779 17:45:21 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:24:14.779 [2024-10-17 17:45:22.049909] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:24:14.779 [2024-10-17 17:45:22.049996] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:14.779 [2024-10-17 17:45:22.141114] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:24:14.779 [2024-10-17 17:45:22.193193] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:14.779 [2024-10-17 17:45:22.193244] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:14.779 [2024-10-17 17:45:22.193252] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:14.779 [2024-10-17 17:45:22.193259] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:14.779 [2024-10-17 17:45:22.193266] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:14.779 [2024-10-17 17:45:22.195187] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:24:14.779 [2024-10-17 17:45:22.195345] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:14.779 [2024-10-17 17:45:22.195345] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:24:15.039 17:45:22 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:15.039 17:45:22 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # return 0 00:24:15.039 17:45:22 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:24:15.039 17:45:22 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@730 -- # xtrace_disable 00:24:15.039 17:45:22 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:24:15.039 17:45:22 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:15.039 17:45:22 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:24:15.299 [2024-10-17 17:45:23.079126] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:15.299 17:45:23 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:24:15.559 Malloc0 00:24:15.559 17:45:23 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:24:15.819 17:45:23 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:24:15.819 17:45:23 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:24:16.080 [2024-10-17 17:45:23.901022] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:16.080 17:45:23 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:24:16.340 [2024-10-17 17:45:24.105532] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:24:16.340 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:24:16.599 [2024-10-17 17:45:24.290056] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@31 -- # bdevperf_pid=156645 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; cat $testdir/try.txt; rm -f $testdir/try.txt; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 15 -f 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@34 -- # waitforlisten 156645 /var/tmp/bdevperf.sock 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@831 -- # '[' -z 156645 ']' 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:24:16.599 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:16.599 17:45:24 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:24:17.536 17:45:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:17.536 17:45:25 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # return 0 00:24:17.536 17:45:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:24:17.536 NVMe0n1 00:24:17.796 17:45:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:24:17.796 00:24:17.796 17:45:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:24:17.796 17:45:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@39 -- # run_test_pid=156821 00:24:17.796 17:45:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@41 -- # sleep 1 00:24:19.174 17:45:26 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:24:19.174 [2024-10-17 17:45:26.857831] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857877] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857882] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857887] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857892] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857897] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857901] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857906] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857915] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857919] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857924] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857929] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857933] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857938] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857942] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857947] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857951] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857956] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857960] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857965] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857970] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857975] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857979] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857984] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857988] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857993] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.857997] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.858002] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.858006] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.858011] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.858015] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 [2024-10-17 17:45:26.858020] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xddcba0 is same with the state(6) to be set 00:24:19.174 17:45:26 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@45 -- # sleep 3 00:24:22.471 17:45:29 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:24:22.471 00:24:22.471 17:45:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:24:22.731 17:45:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@50 -- # sleep 3 00:24:26.031 17:45:33 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:24:26.031 [2024-10-17 17:45:33.618160] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:26.031 17:45:33 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@55 -- # sleep 1 00:24:26.973 17:45:34 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:24:26.973 [2024-10-17 17:45:34.810032] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810074] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810080] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810085] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810090] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810094] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810099] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810104] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810108] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810113] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810117] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810122] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810126] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810131] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810135] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810139] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810144] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810148] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810153] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810158] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810162] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810167] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810176] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810181] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810185] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810190] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810195] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810199] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810204] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 [2024-10-17 17:45:34.810208] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xde83d0 is same with the state(6) to be set 00:24:26.973 17:45:34 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@59 -- # wait 156821 00:24:33.559 { 00:24:33.559 "results": [ 00:24:33.559 { 00:24:33.559 "job": "NVMe0n1", 00:24:33.559 "core_mask": "0x1", 00:24:33.559 "workload": "verify", 00:24:33.559 "status": "finished", 00:24:33.559 "verify_range": { 00:24:33.559 "start": 0, 00:24:33.559 "length": 16384 00:24:33.559 }, 00:24:33.559 "queue_depth": 128, 00:24:33.559 "io_size": 4096, 00:24:33.559 "runtime": 15.00977, 00:24:33.559 "iops": 12530.43850771864, 00:24:33.559 "mibps": 48.94702542077594, 00:24:33.559 "io_failed": 4349, 00:24:33.559 "io_timeout": 0, 00:24:33.559 "avg_latency_us": 9963.071178276203, 00:24:33.559 "min_latency_us": 535.8933333333333, 00:24:33.559 "max_latency_us": 19660.8 00:24:33.559 } 00:24:33.559 ], 00:24:33.559 "core_count": 1 00:24:33.559 } 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@61 -- # killprocess 156645 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@950 -- # '[' -z 156645 ']' 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # kill -0 156645 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # uname 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 156645 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@968 -- # echo 'killing process with pid 156645' 00:24:33.559 killing process with pid 156645 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@969 -- # kill 156645 00:24:33.559 17:45:40 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@974 -- # wait 156645 00:24:33.559 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@63 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:24:33.559 [2024-10-17 17:45:24.370585] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:24:33.559 [2024-10-17 17:45:24.370643] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid156645 ] 00:24:33.559 [2024-10-17 17:45:24.450137] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:33.559 [2024-10-17 17:45:24.485943] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:33.559 Running I/O for 15 seconds... 00:24:33.559 11318.00 IOPS, 44.21 MiB/s [2024-10-17T15:45:41.478Z] [2024-10-17 17:45:26.859423] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.559 [2024-10-17 17:45:26.859456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859467] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.559 [2024-10-17 17:45:26.859475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859483] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.559 [2024-10-17 17:45:26.859490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859498] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.559 [2024-10-17 17:45:26.859506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859513] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1577ca0 is same with the state(6) to be set 00:24:33.559 [2024-10-17 17:45:26.859567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:98648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:98656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:98664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:98672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:98680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:98688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:98696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:98704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:98712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:98720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:98728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:98736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:98744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:98752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:98760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.559 [2024-10-17 17:45:26.859827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.559 [2024-10-17 17:45:26.859836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:98768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:98776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:98784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:98792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:98800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:98808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:98816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:98824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:98832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.859984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:98840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.859991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:98848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:98856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:98864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:98872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:98880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:98888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:98896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:98904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:98912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:98920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:98928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:98936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:98944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:98952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:98960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:98968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:98976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:98984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:98992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:99000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:99008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860353] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:99016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:99024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:99032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:99040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:99048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.560 [2024-10-17 17:45:26.860427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.560 [2024-10-17 17:45:26.860436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:99056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:99064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:99072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:99080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:99088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:99096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:99104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:99112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:99120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:99128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:99136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:99144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:99152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:99160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:99168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:99176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:99184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:99192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:99200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:99208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:99216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:99224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:99232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:99240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:99248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:99256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:99264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:99272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:99280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:99288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:99296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:99304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:99312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.860986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:99320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.860993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.861002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:99328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.861009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.861018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:99336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.861025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.861034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:99344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.561 [2024-10-17 17:45:26.861042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.561 [2024-10-17 17:45:26.861051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:99352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.562 [2024-10-17 17:45:26.861058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:99360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.562 [2024-10-17 17:45:26.861074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:99368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.562 [2024-10-17 17:45:26.861090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:99376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.562 [2024-10-17 17:45:26.861107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:99384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.562 [2024-10-17 17:45:26.861124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:99392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.562 [2024-10-17 17:45:26.861140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:98384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:98392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:98400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:98408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:98416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:98424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:98432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:98440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:98448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:98456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:98464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:98472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:98480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:98488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:98496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:98504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:98512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:98520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:98528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:98536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:98544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:98552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:98560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:98568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:98576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:98584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:98592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:98600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:98608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:98616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:98624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.562 [2024-10-17 17:45:26.861657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.562 [2024-10-17 17:45:26.861667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:99400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:26.861674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:26.861683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:98632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.563 [2024-10-17 17:45:26.861693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:26.861712] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.563 [2024-10-17 17:45:26.861718] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.563 [2024-10-17 17:45:26.861725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:98640 len:8 PRP1 0x0 PRP2 0x0 00:24:33.563 [2024-10-17 17:45:26.861733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:26.861768] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1598780 was disconnected and freed. reset controller. 00:24:33.563 [2024-10-17 17:45:26.861778] bdev_nvme.c:1988:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:24:33.563 [2024-10-17 17:45:26.861786] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:24:33.563 [2024-10-17 17:45:26.865291] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:24:33.563 [2024-10-17 17:45:26.865314] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1577ca0 (9): Bad file descriptor 00:24:33.563 [2024-10-17 17:45:26.908164] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:24:33.563 11112.50 IOPS, 43.41 MiB/s [2024-10-17T15:45:41.482Z] 11093.00 IOPS, 43.33 MiB/s [2024-10-17T15:45:41.482Z] 11443.25 IOPS, 44.70 MiB/s [2024-10-17T15:45:41.482Z] [2024-10-17 17:45:30.428768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:49664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.563 [2024-10-17 17:45:30.428807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:49672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:49680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:49688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:49696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:49704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:49712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:49720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:49728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:49736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:49744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:49752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:49760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:49768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:49776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.428992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:49784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.428997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:49792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:49800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:49808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:49816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:49824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:49832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:49840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:49848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:49856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:49864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:49872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.563 [2024-10-17 17:45:30.429129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.563 [2024-10-17 17:45:30.429136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:49880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:49888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:49896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:49904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:49912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:49920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:49928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:49936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:49944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:49952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:49960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:49968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:49976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:49984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:49992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:50000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:50008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:50016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:50024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:50032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:50040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:50048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:50056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:50064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:50072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:50080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:50088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:50096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:50104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:50112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:50120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:50128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:50136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.564 [2024-10-17 17:45:30.429522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:50144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.564 [2024-10-17 17:45:30.429526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:50152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:50160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:50168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:50176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:50184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:50192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:50200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:50208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:50216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:50224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:50232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:50240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:50248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:50256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:50264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:50272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:50280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:50288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:50296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:50304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:50312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:50320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:50328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429802] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:50336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:50344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:50352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:50360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:50368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:50376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:50384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:50392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:50400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:50408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:50416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:50424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:50432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:50440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:50448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:50456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.565 [2024-10-17 17:45:30.429986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:50464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.565 [2024-10-17 17:45:30.429991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.429998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:50472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:50480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:50488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:50496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:50504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:50512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:50520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:50528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:50536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:50544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:50552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:50560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:50568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:50576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:50584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:50592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:50600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:50608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:50616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:50624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:50632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:50640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:50648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:50656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:50664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:50672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.566 [2024-10-17 17:45:30.430290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430306] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.566 [2024-10-17 17:45:30.430310] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.566 [2024-10-17 17:45:30.430316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:50680 len:8 PRP1 0x0 PRP2 0x0 00:24:33.566 [2024-10-17 17:45:30.430322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430351] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x159a710 was disconnected and freed. reset controller. 00:24:33.566 [2024-10-17 17:45:30.430358] bdev_nvme.c:1988:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4421 to 10.0.0.2:4422 00:24:33.566 [2024-10-17 17:45:30.430373] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.566 [2024-10-17 17:45:30.430379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430385] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.566 [2024-10-17 17:45:30.430390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430396] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.566 [2024-10-17 17:45:30.430401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430407] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.566 [2024-10-17 17:45:30.430412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:30.430417] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:24:33.566 [2024-10-17 17:45:30.430434] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1577ca0 (9): Bad file descriptor 00:24:33.566 [2024-10-17 17:45:30.432820] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:24:33.566 [2024-10-17 17:45:30.460503] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:24:33.566 11662.00 IOPS, 45.55 MiB/s [2024-10-17T15:45:41.485Z] 11882.67 IOPS, 46.42 MiB/s [2024-10-17T15:45:41.485Z] 12031.86 IOPS, 47.00 MiB/s [2024-10-17T15:45:41.485Z] 12195.50 IOPS, 47.64 MiB/s [2024-10-17T15:45:41.485Z] 12274.78 IOPS, 47.95 MiB/s [2024-10-17T15:45:41.485Z] [2024-10-17 17:45:34.810877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:121840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.566 [2024-10-17 17:45:34.810905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:34.810918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:121848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.566 [2024-10-17 17:45:34.810925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:34.810932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:121856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.566 [2024-10-17 17:45:34.810938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.566 [2024-10-17 17:45:34.810945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:121864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.566 [2024-10-17 17:45:34.810950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.810957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:121872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.810966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.810973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:121880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.810978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.810985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:121888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.810990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.810996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:121896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:121904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:121912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:121920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:121928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:121936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:121944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:121952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:121960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:122152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:121968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:121976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:121984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:121992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:122000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:122008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:122016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:122024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:122160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:122168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:122176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:122184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:122192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:122200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:122208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:122216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:122032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:122040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:122048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:122056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:122064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:122072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:122080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.567 [2024-10-17 17:45:34.811372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:122224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:122232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.567 [2024-10-17 17:45:34.811395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.567 [2024-10-17 17:45:34.811401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:122240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:122248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:122256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:122264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:122272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:122280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:122288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:122296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:122304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:122312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:122320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:122328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:122336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:122344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:122088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:122096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:122104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:122112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:122120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:122128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:122136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:122144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:33.568 [2024-10-17 17:45:34.811648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:122352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:122360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:122368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:122376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:122384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:122392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:122400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:122408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:122416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:122424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:122432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:122440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:122448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:122456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.568 [2024-10-17 17:45:34.811811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.568 [2024-10-17 17:45:34.811817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:122464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:122472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:122480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:122488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:122496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:122504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:122512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:122520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:122528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:122536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:122544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:122552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:122560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:122568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.811982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:122576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:33.569 [2024-10-17 17:45:34.811989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812008] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122584 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812026] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812030] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122592 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812045] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812049] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122600 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812063] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812067] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122608 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812082] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812085] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122616 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812100] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812104] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122624 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812118] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812122] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122632 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812136] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812140] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122640 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812156] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812160] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122648 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812174] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812177] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122656 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812192] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812196] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122664 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812210] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812214] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122672 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812228] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812232] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122680 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.569 [2024-10-17 17:45:34.812246] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.569 [2024-10-17 17:45:34.812250] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.569 [2024-10-17 17:45:34.812254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122688 len:8 PRP1 0x0 PRP2 0x0 00:24:33.569 [2024-10-17 17:45:34.812259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812264] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812268] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122696 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812284] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812288] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122704 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812304] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812308] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122712 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812323] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812326] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122720 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812342] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812346] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122728 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812361] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812365] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122736 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812379] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812383] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122744 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812398] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812402] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122752 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812417] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812421] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122760 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.812431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.812438] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.812441] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.812446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122768 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823550] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823558] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122776 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823580] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823585] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122784 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823604] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823609] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122792 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823628] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823633] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122800 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823652] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823657] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122808 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823677] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823682] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122816 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823708] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823713] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122824 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823739] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823744] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122832 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823763] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823768] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122840 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823787] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823792] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122848 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.570 [2024-10-17 17:45:34.823812] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:24:33.570 [2024-10-17 17:45:34.823818] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:24:33.570 [2024-10-17 17:45:34.823823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122856 len:8 PRP1 0x0 PRP2 0x0 00:24:33.570 [2024-10-17 17:45:34.823830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.571 [2024-10-17 17:45:34.823869] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x159a8f0 was disconnected and freed. reset controller. 00:24:33.571 [2024-10-17 17:45:34.823878] bdev_nvme.c:1988:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4422 to 10.0.0.2:4420 00:24:33.571 [2024-10-17 17:45:34.823903] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.571 [2024-10-17 17:45:34.823912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.571 [2024-10-17 17:45:34.823921] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.571 [2024-10-17 17:45:34.823927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.571 [2024-10-17 17:45:34.823935] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.571 [2024-10-17 17:45:34.823942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.571 [2024-10-17 17:45:34.823949] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:33.571 [2024-10-17 17:45:34.823955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:33.571 [2024-10-17 17:45:34.823963] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:24:33.571 [2024-10-17 17:45:34.823999] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1577ca0 (9): Bad file descriptor 00:24:33.571 [2024-10-17 17:45:34.827254] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:24:33.571 [2024-10-17 17:45:34.859983] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:24:33.571 12298.90 IOPS, 48.04 MiB/s [2024-10-17T15:45:41.490Z] 12357.91 IOPS, 48.27 MiB/s [2024-10-17T15:45:41.490Z] 12410.42 IOPS, 48.48 MiB/s [2024-10-17T15:45:41.490Z] 12465.77 IOPS, 48.69 MiB/s [2024-10-17T15:45:41.490Z] 12509.14 IOPS, 48.86 MiB/s 00:24:33.571 Latency(us) 00:24:33.571 [2024-10-17T15:45:41.490Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:33.571 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:24:33.571 Verification LBA range: start 0x0 length 0x4000 00:24:33.571 NVMe0n1 : 15.01 12530.44 48.95 289.74 0.00 9963.07 535.89 19660.80 00:24:33.571 [2024-10-17T15:45:41.490Z] =================================================================================================================== 00:24:33.571 [2024-10-17T15:45:41.490Z] Total : 12530.44 48.95 289.74 0.00 9963.07 535.89 19660.80 00:24:33.571 Received shutdown signal, test time was about 15.000000 seconds 00:24:33.571 00:24:33.571 Latency(us) 00:24:33.571 [2024-10-17T15:45:41.490Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:33.571 [2024-10-17T15:45:41.490Z] =================================================================================================================== 00:24:33.571 [2024-10-17T15:45:41.490Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # grep -c 'Resetting controller successful' 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # count=3 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@67 -- # (( count != 3 )) 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@73 -- # bdevperf_pid=159681 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@75 -- # waitforlisten 159681 /var/tmp/bdevperf.sock 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 1 -f 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@831 -- # '[' -z 159681 ']' 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:24:33.571 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:33.571 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:24:34.142 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:34.142 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # return 0 00:24:34.142 17:45:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:24:34.142 [2024-10-17 17:45:42.009400] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:24:34.142 17:45:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:24:34.402 [2024-10-17 17:45:42.185792] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:24:34.402 17:45:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:24:34.663 NVMe0n1 00:24:34.663 17:45:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:24:34.923 00:24:34.923 17:45:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:24:35.183 00:24:35.183 17:45:43 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:35.183 17:45:43 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # grep -q NVMe0 00:24:35.444 17:45:43 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:24:35.704 17:45:43 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@87 -- # sleep 3 00:24:39.092 17:45:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:39.092 17:45:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # grep -q NVMe0 00:24:39.092 17:45:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@90 -- # run_test_pid=160869 00:24:39.092 17:45:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:24:39.092 17:45:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@92 -- # wait 160869 00:24:40.035 { 00:24:40.035 "results": [ 00:24:40.035 { 00:24:40.035 "job": "NVMe0n1", 00:24:40.035 "core_mask": "0x1", 00:24:40.035 "workload": "verify", 00:24:40.035 "status": "finished", 00:24:40.035 "verify_range": { 00:24:40.035 "start": 0, 00:24:40.035 "length": 16384 00:24:40.035 }, 00:24:40.035 "queue_depth": 128, 00:24:40.035 "io_size": 4096, 00:24:40.035 "runtime": 1.005797, 00:24:40.035 "iops": 12820.678526581407, 00:24:40.035 "mibps": 50.08077549445862, 00:24:40.035 "io_failed": 0, 00:24:40.035 "io_timeout": 0, 00:24:40.035 "avg_latency_us": 9936.115658265478, 00:24:40.035 "min_latency_us": 1331.2, 00:24:40.035 "max_latency_us": 8519.68 00:24:40.035 } 00:24:40.035 ], 00:24:40.035 "core_count": 1 00:24:40.035 } 00:24:40.035 17:45:47 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@94 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:24:40.035 [2024-10-17 17:45:41.066837] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:24:40.035 [2024-10-17 17:45:41.066898] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid159681 ] 00:24:40.035 [2024-10-17 17:45:41.144745] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:40.035 [2024-10-17 17:45:41.174135] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:40.035 [2024-10-17 17:45:43.430831] bdev_nvme.c:1988:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:24:40.035 [2024-10-17 17:45:43.430869] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:40.035 [2024-10-17 17:45:43.430878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:40.035 [2024-10-17 17:45:43.430885] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:40.035 [2024-10-17 17:45:43.430890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:40.035 [2024-10-17 17:45:43.430896] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:40.035 [2024-10-17 17:45:43.430901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:40.035 [2024-10-17 17:45:43.430907] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:40.035 [2024-10-17 17:45:43.430912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:40.035 [2024-10-17 17:45:43.430917] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:24:40.035 [2024-10-17 17:45:43.430939] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:24:40.035 [2024-10-17 17:45:43.430951] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1c9eca0 (9): Bad file descriptor 00:24:40.035 [2024-10-17 17:45:43.523898] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:24:40.035 Running I/O for 1 seconds... 00:24:40.035 12745.00 IOPS, 49.79 MiB/s 00:24:40.035 Latency(us) 00:24:40.035 [2024-10-17T15:45:47.954Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:40.035 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:24:40.035 Verification LBA range: start 0x0 length 0x4000 00:24:40.035 NVMe0n1 : 1.01 12820.68 50.08 0.00 0.00 9936.12 1331.20 8519.68 00:24:40.035 [2024-10-17T15:45:47.954Z] =================================================================================================================== 00:24:40.035 [2024-10-17T15:45:47.954Z] Total : 12820.68 50.08 0.00 0.00 9936.12 1331.20 8519.68 00:24:40.035 17:45:47 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:40.035 17:45:47 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # grep -q NVMe0 00:24:40.295 17:45:47 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@98 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:24:40.295 17:45:48 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:40.295 17:45:48 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # grep -q NVMe0 00:24:40.556 17:45:48 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:24:40.816 17:45:48 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@101 -- # sleep 3 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # grep -q NVMe0 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@108 -- # killprocess 159681 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@950 -- # '[' -z 159681 ']' 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # kill -0 159681 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # uname 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 159681 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@968 -- # echo 'killing process with pid 159681' 00:24:44.118 killing process with pid 159681 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@969 -- # kill 159681 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@974 -- # wait 159681 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@110 -- # sync 00:24:44.118 17:45:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@113 -- # trap - SIGINT SIGTERM EXIT 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@115 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@116 -- # nvmftestfini 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@514 -- # nvmfcleanup 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@121 -- # sync 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@124 -- # set +e 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:44.378 rmmod nvme_tcp 00:24:44.378 rmmod nvme_fabrics 00:24:44.378 rmmod nvme_keyring 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@128 -- # set -e 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@129 -- # return 0 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@515 -- # '[' -n 156043 ']' 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@516 -- # killprocess 156043 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@950 -- # '[' -z 156043 ']' 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # kill -0 156043 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # uname 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 156043 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@968 -- # echo 'killing process with pid 156043' 00:24:44.378 killing process with pid 156043 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@969 -- # kill 156043 00:24:44.378 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@974 -- # wait 156043 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@297 -- # iptr 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@789 -- # iptables-save 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@789 -- # iptables-restore 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:44.639 17:45:52 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:46.552 17:45:54 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:46.552 00:24:46.552 real 0m40.311s 00:24:46.552 user 2m3.324s 00:24:46.552 sys 0m8.831s 00:24:46.552 17:45:54 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:46.552 17:45:54 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:24:46.552 ************************************ 00:24:46.552 END TEST nvmf_failover 00:24:46.552 ************************************ 00:24:46.552 17:45:54 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@26 -- # run_test nvmf_host_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:24:46.552 17:45:54 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:24:46.552 17:45:54 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:46.552 17:45:54 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:24:46.813 ************************************ 00:24:46.813 START TEST nvmf_host_discovery 00:24:46.813 ************************************ 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:24:46.813 * Looking for test storage... 00:24:46.813 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1691 -- # lcov --version 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@344 -- # case "$op" in 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@345 -- # : 1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # decimal 1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # decimal 2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # return 0 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:24:46.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:46.813 --rc genhtml_branch_coverage=1 00:24:46.813 --rc genhtml_function_coverage=1 00:24:46.813 --rc genhtml_legend=1 00:24:46.813 --rc geninfo_all_blocks=1 00:24:46.813 --rc geninfo_unexecuted_blocks=1 00:24:46.813 00:24:46.813 ' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:24:46.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:46.813 --rc genhtml_branch_coverage=1 00:24:46.813 --rc genhtml_function_coverage=1 00:24:46.813 --rc genhtml_legend=1 00:24:46.813 --rc geninfo_all_blocks=1 00:24:46.813 --rc geninfo_unexecuted_blocks=1 00:24:46.813 00:24:46.813 ' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:24:46.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:46.813 --rc genhtml_branch_coverage=1 00:24:46.813 --rc genhtml_function_coverage=1 00:24:46.813 --rc genhtml_legend=1 00:24:46.813 --rc geninfo_all_blocks=1 00:24:46.813 --rc geninfo_unexecuted_blocks=1 00:24:46.813 00:24:46.813 ' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:24:46.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:46.813 --rc genhtml_branch_coverage=1 00:24:46.813 --rc genhtml_function_coverage=1 00:24:46.813 --rc genhtml_legend=1 00:24:46.813 --rc geninfo_all_blocks=1 00:24:46.813 --rc geninfo_unexecuted_blocks=1 00:24:46.813 00:24:46.813 ' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # uname -s 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:46.813 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@5 -- # export PATH 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@51 -- # : 0 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:46.814 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@11 -- # '[' tcp == rdma ']' 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@16 -- # DISCOVERY_PORT=8009 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@17 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@20 -- # NQN=nqn.2016-06.io.spdk:cnode 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@22 -- # HOST_NQN=nqn.2021-12.io.spdk:test 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@23 -- # HOST_SOCK=/tmp/host.sock 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@25 -- # nvmftestinit 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@474 -- # prepare_net_devs 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@436 -- # local -g is_hw=no 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@438 -- # remove_spdk_ns 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:46.814 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:47.075 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:24:47.075 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:24:47.075 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:24:47.075 17:45:54 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # e810=() 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # x722=() 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # mlx=() 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:24:55.213 Found 0000:31:00.0 (0x8086 - 0x159b) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:24:55.213 Found 0000:31:00.1 (0x8086 - 0x159b) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:24:55.213 Found net devices under 0000:31:00.0: cvl_0_0 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:24:55.213 Found net devices under 0000:31:00.1: cvl_0_1 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # is_hw=yes 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:55.213 17:46:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:55.213 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:55.213 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.633 ms 00:24:55.213 00:24:55.213 --- 10.0.0.2 ping statistics --- 00:24:55.213 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:55.213 rtt min/avg/max/mdev = 0.633/0.633/0.633/0.000 ms 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:55.213 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:55.213 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.285 ms 00:24:55.213 00:24:55.213 --- 10.0.0.1 ping statistics --- 00:24:55.213 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:55.213 rtt min/avg/max/mdev = 0.285/0.285/0.285/0.000 ms 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@448 -- # return 0 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:24:55.213 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@30 -- # nvmfappstart -m 0x2 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@724 -- # xtrace_disable 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@507 -- # nvmfpid=166085 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@508 -- # waitforlisten 166085 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@831 -- # '[' -z 166085 ']' 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:55.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:55.214 17:46:02 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.214 [2024-10-17 17:46:02.423657] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:24:55.214 [2024-10-17 17:46:02.423741] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:55.214 [2024-10-17 17:46:02.516269] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:55.214 [2024-10-17 17:46:02.566751] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:55.214 [2024-10-17 17:46:02.566803] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:55.214 [2024-10-17 17:46:02.566812] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:55.214 [2024-10-17 17:46:02.566819] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:55.214 [2024-10-17 17:46:02.566825] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:55.214 [2024-10-17 17:46:02.567671] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # return 0 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@730 -- # xtrace_disable 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@32 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.475 [2024-10-17 17:46:03.287593] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2014-08.org.nvmexpress.discovery -t tcp -a 10.0.0.2 -s 8009 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.475 [2024-10-17 17:46:03.299848] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@35 -- # rpc_cmd bdev_null_create null0 1000 512 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.475 null0 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@36 -- # rpc_cmd bdev_null_create null1 1000 512 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.475 null1 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@37 -- # rpc_cmd bdev_wait_for_examine 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@45 -- # hostpid=166423 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@46 -- # waitforlisten 166423 /tmp/host.sock 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@831 -- # '[' -z 166423 ']' 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # local rpc_addr=/tmp/host.sock 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:24:55.475 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:55.475 17:46:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:55.735 [2024-10-17 17:46:03.397525] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:24:55.735 [2024-10-17 17:46:03.397589] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid166423 ] 00:24:55.735 [2024-10-17 17:46:03.479744] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:55.735 [2024-10-17 17:46:03.532881] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # return 0 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@48 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@50 -- # rpc_cmd -s /tmp/host.sock log_set_flag bdev_nvme 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@51 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.306 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.566 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.566 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@72 -- # notify_id=0 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # get_subsystem_names 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # [[ '' == '' ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # get_bdev_list 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # [[ '' == '' ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@86 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # get_subsystem_names 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # [[ '' == '' ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # get_bdev_list 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # [[ '' == '' ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@90 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # get_subsystem_names 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:56.567 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # [[ '' == '' ]] 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # get_bdev_list 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # [[ '' == '' ]] 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@96 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.827 [2024-10-17 17:46:04.575141] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.827 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # get_subsystem_names 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # [[ '' == '' ]] 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # get_bdev_list 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # [[ '' == '' ]] 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@99 -- # is_notification_count_eq 0 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=0 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@103 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2021-12.io.spdk:test 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@105 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:56.828 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:57.089 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ '' == \n\v\m\e\0 ]] 00:24:57.089 17:46:04 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # sleep 1 00:24:57.350 [2024-10-17 17:46:05.247111] bdev_nvme.c:7151:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:24:57.350 [2024-10-17 17:46:05.247145] bdev_nvme.c:7237:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:24:57.350 [2024-10-17 17:46:05.247161] bdev_nvme.c:7114:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:24:57.610 [2024-10-17 17:46:05.335399] bdev_nvme.c:7080:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:24:57.610 [2024-10-17 17:46:05.397479] bdev_nvme.c:6970:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:24:57.610 [2024-10-17 17:46:05.397515] bdev_nvme.c:6929:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:57.871 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@106 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1"' ']]' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 == \n\v\m\e\0\n\1 ]] 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@107 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT"' ']]' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4420 == \4\4\2\0 ]] 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@108 -- # is_notification_count_eq 1 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=1 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@111 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null1 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.132 17:46:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.132 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.132 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@113 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:24:58.132 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:24:58.132 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.132 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.133 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@114 -- # is_notification_count_eq 1 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 1 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@118 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.393 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.393 [2024-10-17 17:46:06.307610] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:24:58.393 [2024-10-17 17:46:06.308407] bdev_nvme.c:7133:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:24:58.393 [2024-10-17 17:46:06.308436] bdev_nvme.c:7114:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@120 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@121 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:58.653 [2024-10-17 17:46:06.396141] bdev_nvme.c:7075:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new path for nvme0 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@122 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:24:58.653 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:58.653 [2024-10-17 17:46:06.460927] bdev_nvme.c:6970:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:24:58.653 [2024-10-17 17:46:06.460945] bdev_nvme.c:6929:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:24:58.654 [2024-10-17 17:46:06.460951] bdev_nvme.c:6929:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:24:58.654 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4420 == \4\4\2\0\ \4\4\2\1 ]] 00:24:58.654 17:46:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # sleep 1 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:24:59.593 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4420 4421 == \4\4\2\0\ \4\4\2\1 ]] 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@123 -- # is_notification_count_eq 0 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@127 -- # rpc_cmd nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:59.854 [2024-10-17 17:46:07.563023] bdev_nvme.c:7133:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:24:59.854 [2024-10-17 17:46:07.563041] bdev_nvme.c:7114:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:24:59.854 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:59.854 [2024-10-17 17:46:07.568347] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:59.854 [2024-10-17 17:46:07.568361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:59.854 [2024-10-17 17:46:07.568367] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:59.855 [2024-10-17 17:46:07.568373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:59.855 [2024-10-17 17:46:07.568378] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:59.855 [2024-10-17 17:46:07.568383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@129 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:24:59.855 [2024-10-17 17:46:07.568389] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:59.855 [2024-10-17 17:46:07.568401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:59.855 [2024-10-17 17:46:07.568406] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:24:59.855 [2024-10-17 17:46:07.578364] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.588398] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:59.855 [2024-10-17 17:46:07.588483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:24:59.855 [2024-10-17 17:46:07.588493] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x22fdcd0 with addr=10.0.0.2, port=4420 00:24:59.855 [2024-10-17 17:46:07.588498] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 [2024-10-17 17:46:07.588507] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.588514] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:59.855 [2024-10-17 17:46:07.588519] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:59.855 [2024-10-17 17:46:07.588525] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:59.855 [2024-10-17 17:46:07.588534] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:59.855 [2024-10-17 17:46:07.598446] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:59.855 [2024-10-17 17:46:07.598893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:24:59.855 [2024-10-17 17:46:07.598923] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x22fdcd0 with addr=10.0.0.2, port=4420 00:24:59.855 [2024-10-17 17:46:07.598932] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 [2024-10-17 17:46:07.598947] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.598956] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:59.855 [2024-10-17 17:46:07.598960] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:59.855 [2024-10-17 17:46:07.598966] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:59.855 [2024-10-17 17:46:07.598982] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@130 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:24:59.855 [2024-10-17 17:46:07.608490] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:59.855 [2024-10-17 17:46:07.608921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:24:59.855 [2024-10-17 17:46:07.608951] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x22fdcd0 with addr=10.0.0.2, port=4420 00:24:59.855 [2024-10-17 17:46:07.608960] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 [2024-10-17 17:46:07.608974] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.608983] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:59.855 [2024-10-17 17:46:07.608988] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:59.855 [2024-10-17 17:46:07.608993] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:59.855 [2024-10-17 17:46:07.609005] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:59.855 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:24:59.855 [2024-10-17 17:46:07.618541] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:59.855 [2024-10-17 17:46:07.618908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:24:59.855 [2024-10-17 17:46:07.618938] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x22fdcd0 with addr=10.0.0.2, port=4420 00:24:59.855 [2024-10-17 17:46:07.618947] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 [2024-10-17 17:46:07.618961] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.618969] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:59.855 [2024-10-17 17:46:07.618974] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:59.855 [2024-10-17 17:46:07.618980] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:59.855 [2024-10-17 17:46:07.618991] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:59.855 [2024-10-17 17:46:07.628591] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:59.855 [2024-10-17 17:46:07.628803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:24:59.855 [2024-10-17 17:46:07.628813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x22fdcd0 with addr=10.0.0.2, port=4420 00:24:59.855 [2024-10-17 17:46:07.628818] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 [2024-10-17 17:46:07.628826] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.628834] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:59.855 [2024-10-17 17:46:07.628838] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:59.855 [2024-10-17 17:46:07.628843] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:59.855 [2024-10-17 17:46:07.628851] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:59.855 [2024-10-17 17:46:07.638637] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:59.855 [2024-10-17 17:46:07.638820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:24:59.855 [2024-10-17 17:46:07.638829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x22fdcd0 with addr=10.0.0.2, port=4420 00:24:59.855 [2024-10-17 17:46:07.638835] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 [2024-10-17 17:46:07.638843] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.638850] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:59.855 [2024-10-17 17:46:07.638854] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:59.855 [2024-10-17 17:46:07.638859] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:59.855 [2024-10-17 17:46:07.638867] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:59.855 [2024-10-17 17:46:07.648680] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:59.855 [2024-10-17 17:46:07.648989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:24:59.855 [2024-10-17 17:46:07.648998] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x22fdcd0 with addr=10.0.0.2, port=4420 00:24:59.855 [2024-10-17 17:46:07.649003] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22fdcd0 is same with the state(6) to be set 00:24:59.855 [2024-10-17 17:46:07.649011] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22fdcd0 (9): Bad file descriptor 00:24:59.855 [2024-10-17 17:46:07.649018] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:59.855 [2024-10-17 17:46:07.649022] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:59.855 [2024-10-17 17:46:07.649026] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:59.855 [2024-10-17 17:46:07.649034] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:59.855 [2024-10-17 17:46:07.650280] bdev_nvme.c:6938:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 not found 00:24:59.855 [2024-10-17 17:46:07.650292] bdev_nvme.c:6929:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@131 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_SECOND_PORT"' ']]' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4421 == \4\4\2\1 ]] 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@132 -- # is_notification_count_eq 0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@134 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_stop_discovery -b nvme 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:59.856 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@136 -- # waitforcondition '[[ "$(get_subsystem_names)" == "" ]]' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "" ]]' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '""' ']]' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ '' == '' ]] 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@137 -- # waitforcondition '[[ "$(get_bdev_list)" == "" ]]' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "" ]]' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '""' ']]' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ '' == '' ]] 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@138 -- # is_notification_count_eq 2 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=2 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=2 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=4 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@141 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:00.116 17:46:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:01.498 [2024-10-17 17:46:08.997608] bdev_nvme.c:7151:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:25:01.498 [2024-10-17 17:46:08.997622] bdev_nvme.c:7237:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:25:01.498 [2024-10-17 17:46:08.997631] bdev_nvme.c:7114:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:25:01.498 [2024-10-17 17:46:09.086897] bdev_nvme.c:7080:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new subsystem nvme0 00:25:01.498 [2024-10-17 17:46:09.149494] bdev_nvme.c:6970:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:25:01.498 [2024-10-17 17:46:09.149518] bdev_nvme.c:6929:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@143 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@650 -- # local es=0 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:01.498 request: 00:25:01.498 { 00:25:01.498 "name": "nvme", 00:25:01.498 "trtype": "tcp", 00:25:01.498 "traddr": "10.0.0.2", 00:25:01.498 "adrfam": "ipv4", 00:25:01.498 "trsvcid": "8009", 00:25:01.498 "hostnqn": "nqn.2021-12.io.spdk:test", 00:25:01.498 "wait_for_attach": true, 00:25:01.498 "method": "bdev_nvme_start_discovery", 00:25:01.498 "req_id": 1 00:25:01.498 } 00:25:01.498 Got JSON-RPC error response 00:25:01.498 response: 00:25:01.498 { 00:25:01.498 "code": -17, 00:25:01.498 "message": "File exists" 00:25:01.498 } 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # es=1 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # get_discovery_ctrlrs 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # [[ nvme == \n\v\m\e ]] 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # get_bdev_list 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@149 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@650 -- # local es=0 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:25:01.498 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:01.499 request: 00:25:01.499 { 00:25:01.499 "name": "nvme_second", 00:25:01.499 "trtype": "tcp", 00:25:01.499 "traddr": "10.0.0.2", 00:25:01.499 "adrfam": "ipv4", 00:25:01.499 "trsvcid": "8009", 00:25:01.499 "hostnqn": "nqn.2021-12.io.spdk:test", 00:25:01.499 "wait_for_attach": true, 00:25:01.499 "method": "bdev_nvme_start_discovery", 00:25:01.499 "req_id": 1 00:25:01.499 } 00:25:01.499 Got JSON-RPC error response 00:25:01.499 response: 00:25:01.499 { 00:25:01.499 "code": -17, 00:25:01.499 "message": "File exists" 00:25:01.499 } 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # es=1 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # get_discovery_ctrlrs 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # [[ nvme == \n\v\m\e ]] 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # get_bdev_list 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@155 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@650 -- # local es=0 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.499 17:46:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:02.883 [2024-10-17 17:46:10.408929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:02.883 [2024-10-17 17:46:10.408964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x23394f0 with addr=10.0.0.2, port=8010 00:25:02.883 [2024-10-17 17:46:10.408980] nvme_tcp.c:2723:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:25:02.883 [2024-10-17 17:46:10.408986] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:25:02.883 [2024-10-17 17:46:10.408992] bdev_nvme.c:7219:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:25:03.824 [2024-10-17 17:46:11.411247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:03.824 [2024-10-17 17:46:11.411268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x23394f0 with addr=10.0.0.2, port=8010 00:25:03.824 [2024-10-17 17:46:11.411277] nvme_tcp.c:2723:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:25:03.824 [2024-10-17 17:46:11.411282] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:25:03.824 [2024-10-17 17:46:11.411287] bdev_nvme.c:7219:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:25:04.764 [2024-10-17 17:46:12.413286] bdev_nvme.c:7194:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] timed out while attaching discovery ctrlr 00:25:04.764 request: 00:25:04.764 { 00:25:04.764 "name": "nvme_second", 00:25:04.764 "trtype": "tcp", 00:25:04.764 "traddr": "10.0.0.2", 00:25:04.764 "adrfam": "ipv4", 00:25:04.764 "trsvcid": "8010", 00:25:04.764 "hostnqn": "nqn.2021-12.io.spdk:test", 00:25:04.764 "wait_for_attach": false, 00:25:04.764 "attach_timeout_ms": 3000, 00:25:04.764 "method": "bdev_nvme_start_discovery", 00:25:04.764 "req_id": 1 00:25:04.764 } 00:25:04.764 Got JSON-RPC error response 00:25:04.764 response: 00:25:04.764 { 00:25:04.764 "code": -110, 00:25:04.764 "message": "Connection timed out" 00:25:04.764 } 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # es=1 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # get_discovery_ctrlrs 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # [[ nvme == \n\v\m\e ]] 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@159 -- # trap - SIGINT SIGTERM EXIT 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@161 -- # kill 166423 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@162 -- # nvmftestfini 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@514 -- # nvmfcleanup 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@121 -- # sync 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@124 -- # set +e 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:04.764 rmmod nvme_tcp 00:25:04.764 rmmod nvme_fabrics 00:25:04.764 rmmod nvme_keyring 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@128 -- # set -e 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@129 -- # return 0 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@515 -- # '[' -n 166085 ']' 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@516 -- # killprocess 166085 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@950 -- # '[' -z 166085 ']' 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@954 -- # kill -0 166085 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@955 -- # uname 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 166085 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@968 -- # echo 'killing process with pid 166085' 00:25:04.764 killing process with pid 166085 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@969 -- # kill 166085 00:25:04.764 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@974 -- # wait 166085 00:25:05.024 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:25:05.024 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:25:05.024 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@297 -- # iptr 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@789 -- # iptables-save 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@789 -- # iptables-restore 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:05.025 17:46:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:06.936 17:46:14 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:06.936 00:25:06.936 real 0m20.306s 00:25:06.936 user 0m23.301s 00:25:06.936 sys 0m7.323s 00:25:06.936 17:46:14 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:06.936 17:46:14 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:25:06.936 ************************************ 00:25:06.936 END TEST nvmf_host_discovery 00:25:06.936 ************************************ 00:25:06.936 17:46:14 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@27 -- # run_test nvmf_host_multipath_status /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:25:06.936 17:46:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:25:06.936 17:46:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:06.936 17:46:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.197 ************************************ 00:25:07.197 START TEST nvmf_host_multipath_status 00:25:07.197 ************************************ 00:25:07.197 17:46:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:25:07.197 * Looking for test storage... 00:25:07.197 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:07.197 17:46:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:25:07.197 17:46:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1691 -- # lcov --version 00:25:07.197 17:46:14 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # IFS=.-: 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # read -ra ver1 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # IFS=.-: 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # read -ra ver2 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@338 -- # local 'op=<' 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@340 -- # ver1_l=2 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@341 -- # ver2_l=1 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@344 -- # case "$op" in 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@345 -- # : 1 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:07.197 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # decimal 1 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=1 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 1 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # ver1[v]=1 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # decimal 2 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=2 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 2 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # ver2[v]=2 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # return 0 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:25:07.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:07.198 --rc genhtml_branch_coverage=1 00:25:07.198 --rc genhtml_function_coverage=1 00:25:07.198 --rc genhtml_legend=1 00:25:07.198 --rc geninfo_all_blocks=1 00:25:07.198 --rc geninfo_unexecuted_blocks=1 00:25:07.198 00:25:07.198 ' 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:25:07.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:07.198 --rc genhtml_branch_coverage=1 00:25:07.198 --rc genhtml_function_coverage=1 00:25:07.198 --rc genhtml_legend=1 00:25:07.198 --rc geninfo_all_blocks=1 00:25:07.198 --rc geninfo_unexecuted_blocks=1 00:25:07.198 00:25:07.198 ' 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:25:07.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:07.198 --rc genhtml_branch_coverage=1 00:25:07.198 --rc genhtml_function_coverage=1 00:25:07.198 --rc genhtml_legend=1 00:25:07.198 --rc geninfo_all_blocks=1 00:25:07.198 --rc geninfo_unexecuted_blocks=1 00:25:07.198 00:25:07.198 ' 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:25:07.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:07.198 --rc genhtml_branch_coverage=1 00:25:07.198 --rc genhtml_function_coverage=1 00:25:07.198 --rc genhtml_legend=1 00:25:07.198 --rc geninfo_all_blocks=1 00:25:07.198 --rc geninfo_unexecuted_blocks=1 00:25:07.198 00:25:07.198 ' 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # uname -s 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@15 -- # shopt -s extglob 00:25:07.198 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@5 -- # export PATH 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@51 -- # : 0 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:07.459 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@12 -- # MALLOC_BDEV_SIZE=64 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:25:07.459 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@16 -- # bpf_sh=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/bpftrace.sh 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@18 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@21 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@31 -- # nvmftestinit 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@474 -- # prepare_net_devs 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@436 -- # local -g is_hw=no 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@438 -- # remove_spdk_ns 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@309 -- # xtrace_disable 00:25:07.460 17:46:15 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # pci_devs=() 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # net_devs=() 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # e810=() 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # local -ga e810 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # x722=() 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # local -ga x722 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # mlx=() 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # local -ga mlx 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:15.603 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:25:15.604 Found 0000:31:00.0 (0x8086 - 0x159b) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:25:15.604 Found 0000:31:00.1 (0x8086 - 0x159b) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:25:15.604 Found net devices under 0000:31:00.0: cvl_0_0 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:25:15.604 Found net devices under 0000:31:00.1: cvl_0_1 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # is_hw=yes 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:15.604 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:15.604 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.608 ms 00:25:15.604 00:25:15.604 --- 10.0.0.2 ping statistics --- 00:25:15.604 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:15.604 rtt min/avg/max/mdev = 0.608/0.608/0.608/0.000 ms 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:15.604 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:15.604 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.249 ms 00:25:15.604 00:25:15.604 --- 10.0.0.1 ping statistics --- 00:25:15.604 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:15.604 rtt min/avg/max/mdev = 0.249/0.249/0.249/0.000 ms 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@448 -- # return 0 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@33 -- # nvmfappstart -m 0x3 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@724 -- # xtrace_disable 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@507 -- # nvmfpid=172611 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@508 -- # waitforlisten 172611 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@831 -- # '[' -z 172611 ']' 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:15.604 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:15.604 17:46:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:25:15.604 [2024-10-17 17:46:22.873102] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:25:15.604 [2024-10-17 17:46:22.873166] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:15.605 [2024-10-17 17:46:22.963526] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:25:15.605 [2024-10-17 17:46:23.015161] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:15.605 [2024-10-17 17:46:23.015211] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:15.605 [2024-10-17 17:46:23.015219] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:15.605 [2024-10-17 17:46:23.015227] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:15.605 [2024-10-17 17:46:23.015233] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:15.605 [2024-10-17 17:46:23.016984] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:15.605 [2024-10-17 17:46:23.017084] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # return 0 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@730 -- # xtrace_disable 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@34 -- # nvmfapp_pid=172611 00:25:15.866 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:25:16.127 [2024-10-17 17:46:23.900350] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:16.127 17:46:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:25:16.387 Malloc0 00:25:16.387 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -r -m 2 00:25:16.648 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:16.909 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:16.909 [2024-10-17 17:46:24.737930] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:16.909 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:25:17.171 [2024-10-17 17:46:24.938460] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@45 -- # bdevperf_pid=173003 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 90 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@47 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@48 -- # waitforlisten 173003 /var/tmp/bdevperf.sock 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@831 -- # '[' -z 173003 ']' 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:25:17.171 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:17.171 17:46:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:25:18.115 17:46:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:18.115 17:46:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # return 0 00:25:18.115 17:46:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_options -r -1 00:25:18.375 17:46:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:25:18.636 Nvme0n1 00:25:18.636 17:46:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:25:18.897 Nvme0n1 00:25:18.897 17:46:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@78 -- # sleep 2 00:25:18.897 17:46:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 120 -s /var/tmp/bdevperf.sock perform_tests 00:25:20.815 17:46:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@90 -- # set_ANA_state optimized optimized 00:25:20.815 17:46:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:25:21.076 17:46:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:25:21.337 17:46:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@91 -- # sleep 1 00:25:22.279 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@92 -- # check_status true false true true true true 00:25:22.279 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:25:22.279 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:22.279 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:22.539 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:22.799 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:22.799 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:22.799 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:22.799 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:23.059 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:23.059 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:23.059 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:23.059 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:23.319 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:23.319 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:25:23.319 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:23.319 17:46:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:23.319 17:46:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:23.319 17:46:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@94 -- # set_ANA_state non_optimized optimized 00:25:23.319 17:46:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:25:23.580 17:46:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:25:23.841 17:46:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@95 -- # sleep 1 00:25:24.780 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@96 -- # check_status false true true true true true 00:25:24.780 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:25:24.780 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:24.780 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:25.039 17:46:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:25.298 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:25.298 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:25.298 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:25.298 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:25.557 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:25.817 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:25.817 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@100 -- # set_ANA_state non_optimized non_optimized 00:25:25.817 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:25:25.817 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:25:26.078 17:46:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@101 -- # sleep 1 00:25:27.020 17:46:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@102 -- # check_status true false true true true true 00:25:27.020 17:46:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:25:27.020 17:46:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:27.020 17:46:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:27.279 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:27.279 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:25:27.279 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:27.279 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:27.539 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:27.539 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:27.539 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:27.539 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:27.798 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:28.058 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:28.058 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:25:28.058 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:28.058 17:46:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:28.317 17:46:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:28.317 17:46:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@104 -- # set_ANA_state non_optimized inaccessible 00:25:28.317 17:46:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:25:28.317 17:46:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:25:28.577 17:46:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@105 -- # sleep 1 00:25:29.596 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@106 -- # check_status true false true true true false 00:25:29.596 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:25:29.596 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:29.597 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:29.872 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:30.132 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:30.132 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:30.132 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:30.132 17:46:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:30.392 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:30.392 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:30.392 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:30.392 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:30.651 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:30.651 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:25:30.651 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:30.651 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:30.651 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:30.651 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@108 -- # set_ANA_state inaccessible inaccessible 00:25:30.651 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:25:30.911 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:25:31.172 17:46:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@109 -- # sleep 1 00:25:32.113 17:46:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@110 -- # check_status false false true true false false 00:25:32.113 17:46:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:25:32.113 17:46:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:32.113 17:46:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:32.113 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:32.113 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:25:32.113 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:32.113 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:32.373 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:32.373 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:32.373 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:32.373 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:32.632 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:32.632 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:32.632 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:32.632 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:32.891 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:33.151 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:33.151 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@112 -- # set_ANA_state inaccessible optimized 00:25:33.151 17:46:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:25:33.411 17:46:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:25:33.411 17:46:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@113 -- # sleep 1 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@114 -- # check_status false true true true false true 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:34.794 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:35.053 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:35.053 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:35.053 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:35.053 17:46:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:35.313 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:35.573 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:35.573 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_multipath_policy -b Nvme0n1 -p active_active 00:25:35.834 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@119 -- # set_ANA_state optimized optimized 00:25:35.834 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:25:35.834 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:25:36.095 17:46:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@120 -- # sleep 1 00:25:37.036 17:46:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@121 -- # check_status true true true true true true 00:25:37.036 17:46:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:25:37.036 17:46:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:37.036 17:46:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:37.297 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:37.297 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:25:37.297 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:37.297 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:37.558 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:37.818 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:37.818 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:37.818 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:37.818 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:38.078 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:38.078 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:25:38.078 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:38.078 17:46:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:38.338 17:46:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:38.338 17:46:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@123 -- # set_ANA_state non_optimized optimized 00:25:38.338 17:46:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:25:38.338 17:46:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:25:38.601 17:46:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@124 -- # sleep 1 00:25:39.553 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@125 -- # check_status false true true true true true 00:25:39.553 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:25:39.553 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:39.553 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:39.812 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:39.812 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:25:39.812 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:39.812 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:40.071 17:46:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:40.330 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:40.330 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:40.330 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:40.330 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:40.589 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:40.589 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:25:40.589 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:40.589 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:40.589 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:40.589 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@129 -- # set_ANA_state non_optimized non_optimized 00:25:40.589 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:25:40.849 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:25:41.109 17:46:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@130 -- # sleep 1 00:25:42.050 17:46:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@131 -- # check_status true true true true true true 00:25:42.050 17:46:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:25:42.050 17:46:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:42.050 17:46:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:42.309 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:42.309 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:25:42.309 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:42.309 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:42.568 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:42.828 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:42.828 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:42.828 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:42.828 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:43.087 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:43.087 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:25:43.087 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:43.087 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:43.087 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:43.087 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@133 -- # set_ANA_state non_optimized inaccessible 00:25:43.087 17:46:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:25:43.348 17:46:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:25:43.609 17:46:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@134 -- # sleep 1 00:25:44.549 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@135 -- # check_status true false true true true false 00:25:44.549 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:25:44.549 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:44.549 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:44.808 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:25:45.067 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:45.067 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:25:45.067 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:45.067 17:46:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:25:45.325 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:45.325 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:25:45.325 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:45.326 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:25:45.326 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:25:45.326 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:25:45.326 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:25:45.326 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@137 -- # killprocess 173003 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@950 -- # '[' -z 173003 ']' 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # kill -0 173003 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # uname 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 173003 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@968 -- # echo 'killing process with pid 173003' 00:25:45.585 killing process with pid 173003 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@969 -- # kill 173003 00:25:45.585 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@974 -- # wait 173003 00:25:45.585 { 00:25:45.585 "results": [ 00:25:45.585 { 00:25:45.585 "job": "Nvme0n1", 00:25:45.585 "core_mask": "0x4", 00:25:45.585 "workload": "verify", 00:25:45.585 "status": "terminated", 00:25:45.585 "verify_range": { 00:25:45.585 "start": 0, 00:25:45.585 "length": 16384 00:25:45.585 }, 00:25:45.585 "queue_depth": 128, 00:25:45.585 "io_size": 4096, 00:25:45.585 "runtime": 26.645777, 00:25:45.585 "iops": 12102.142864889998, 00:25:45.585 "mibps": 47.273995565976556, 00:25:45.585 "io_failed": 0, 00:25:45.585 "io_timeout": 0, 00:25:45.585 "avg_latency_us": 10557.918765697794, 00:25:45.585 "min_latency_us": 168.96, 00:25:45.585 "max_latency_us": 3019898.88 00:25:45.585 } 00:25:45.585 ], 00:25:45.585 "core_count": 1 00:25:45.585 } 00:25:45.848 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@139 -- # wait 173003 00:25:45.848 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@141 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:25:45.848 [2024-10-17 17:46:25.025342] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:25:45.848 [2024-10-17 17:46:25.025430] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid173003 ] 00:25:45.848 [2024-10-17 17:46:25.108877] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:45.848 [2024-10-17 17:46:25.159791] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:45.848 Running I/O for 90 seconds... 00:25:45.848 10050.00 IOPS, 39.26 MiB/s [2024-10-17T15:46:53.767Z] 10603.50 IOPS, 41.42 MiB/s [2024-10-17T15:46:53.767Z] 10752.67 IOPS, 42.00 MiB/s [2024-10-17T15:46:53.767Z] 11095.00 IOPS, 43.34 MiB/s [2024-10-17T15:46:53.767Z] 11475.60 IOPS, 44.83 MiB/s [2024-10-17T15:46:53.767Z] 11726.00 IOPS, 45.80 MiB/s [2024-10-17T15:46:53.767Z] 11885.14 IOPS, 46.43 MiB/s [2024-10-17T15:46:53.767Z] 12019.25 IOPS, 46.95 MiB/s [2024-10-17T15:46:53.767Z] 12151.11 IOPS, 47.47 MiB/s [2024-10-17T15:46:53.767Z] 12219.80 IOPS, 47.73 MiB/s [2024-10-17T15:46:53.767Z] 12279.18 IOPS, 47.97 MiB/s [2024-10-17T15:46:53.767Z] [2024-10-17 17:46:38.665417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:122120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.848 [2024-10-17 17:46:38.665451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:45.848 [2024-10-17 17:46:38.665823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:122128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.848 [2024-10-17 17:46:38.665837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:25:45.848 [2024-10-17 17:46:38.665850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:122136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.848 [2024-10-17 17:46:38.665856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:25:45.848 [2024-10-17 17:46:38.665867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:122144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.848 [2024-10-17 17:46:38.665873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:25:45.848 [2024-10-17 17:46:38.665884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:122152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.848 [2024-10-17 17:46:38.665890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:25:45.848 [2024-10-17 17:46:38.665900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:122160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.848 [2024-10-17 17:46:38.665906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:25:45.848 [2024-10-17 17:46:38.665917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:122168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.848 [2024-10-17 17:46:38.665923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:25:45.848 [2024-10-17 17:46:38.665933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:122176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.665939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.665949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:122184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.665954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.665965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:122192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.665975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.665985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:122200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.665990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:122208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:122216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:122224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:122232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:122240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:122248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:122256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:122264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:122272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:122280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:122288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:122296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:122304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:122312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:122320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:122328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:122336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:122344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:122352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:122360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:122368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:122376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:122384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:122392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:122400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:122408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:122416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:122424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:122432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:122440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:122448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:122456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:122464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:122472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:122480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:122488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.849 [2024-10-17 17:46:38.666551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:25:45.849 [2024-10-17 17:46:38.666561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:122496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:122504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:122512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:122520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:122528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:122536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:122544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:122552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:122560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666707] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:122568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:122576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:122584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:122592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:122600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:122608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:122616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:122624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:122632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:122640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:122648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.666983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:122656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.666988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:122664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:122672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:122680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:122688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:122696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:122704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:122712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:122720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:122728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:121936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.850 [2024-10-17 17:46:38.667180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:121944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.850 [2024-10-17 17:46:38.667199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:121952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.850 [2024-10-17 17:46:38.667217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:121960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.850 [2024-10-17 17:46:38.667235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:121968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.850 [2024-10-17 17:46:38.667254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:121976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.850 [2024-10-17 17:46:38.667273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:121984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.850 [2024-10-17 17:46:38.667292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:122736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:122744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:122752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:122760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.850 [2024-10-17 17:46:38.667366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:25:45.850 [2024-10-17 17:46:38.667379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:122768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:122776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:122784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:122792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:122800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:122808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:122816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:122824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:122832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:122840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:122848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:122856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:122864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:122872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:122880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:122888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:122896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:122904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:122912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:122920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:122928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:122936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:122944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.667876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:121992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.667897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:122000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.667917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:122008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.667938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:122016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.667958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:122024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.667979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.667994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:122032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.667999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:122040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:122048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:122056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:122064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:122072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:122080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:122088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:122096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:122104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:122112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.851 [2024-10-17 17:46:38.668207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:25:45.851 [2024-10-17 17:46:38.668223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:122952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.851 [2024-10-17 17:46:38.668228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:25:45.851 12192.83 IOPS, 47.63 MiB/s [2024-10-17T15:46:53.770Z] 11254.92 IOPS, 43.96 MiB/s [2024-10-17T15:46:53.770Z] 10451.00 IOPS, 40.82 MiB/s [2024-10-17T15:46:53.770Z] 9861.27 IOPS, 38.52 MiB/s [2024-10-17T15:46:53.770Z] 10054.06 IOPS, 39.27 MiB/s [2024-10-17T15:46:53.770Z] 10217.94 IOPS, 39.91 MiB/s [2024-10-17T15:46:53.770Z] 10665.78 IOPS, 41.66 MiB/s [2024-10-17T15:46:53.770Z] 11050.95 IOPS, 43.17 MiB/s [2024-10-17T15:46:53.770Z] 11266.20 IOPS, 44.01 MiB/s [2024-10-17T15:46:53.771Z] 11333.90 IOPS, 44.27 MiB/s [2024-10-17T15:46:53.771Z] 11393.77 IOPS, 44.51 MiB/s [2024-10-17T15:46:53.771Z] 11669.52 IOPS, 45.58 MiB/s [2024-10-17T15:46:53.771Z] 11931.29 IOPS, 46.61 MiB/s [2024-10-17T15:46:53.771Z] [2024-10-17 17:46:51.277563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:126016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:126032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:126048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:126064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:126080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:126096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:126112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:126128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:126144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:126160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:126176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:126192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:126208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:125688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.852 [2024-10-17 17:46:51.277833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:126216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:126232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:126248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:126264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:126280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:126296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:126312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:126328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:126344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.277986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:126360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.277992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:126376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:126392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:126408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:126424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:125704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.852 [2024-10-17 17:46:51.278826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:126448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:126464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:126480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:126496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:45.852 [2024-10-17 17:46:51.278897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:126512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.852 [2024-10-17 17:46:51.278903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.278913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:126528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.278918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.278928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:126544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.278933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.278943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:126560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.278952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.278963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:126576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.278968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.278979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:126592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.278984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.278995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:126608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.279000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:126624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.279015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:125712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:126648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.279046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:126664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.279061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:126680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:45.853 [2024-10-17 17:46:51.279077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:125728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:125760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:125792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:125824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:125856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:125888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:125920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:125952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:125984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:125720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:125752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:125784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:125816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:125848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:125880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:125912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:125944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:125976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:25:45.853 [2024-10-17 17:46:51.279880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:126008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:45.853 [2024-10-17 17:46:51.279889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:25:45.853 12052.52 IOPS, 47.08 MiB/s [2024-10-17T15:46:53.772Z] 12084.65 IOPS, 47.21 MiB/s [2024-10-17T15:46:53.772Z] Received shutdown signal, test time was about 26.646389 seconds 00:25:45.853 00:25:45.853 Latency(us) 00:25:45.853 [2024-10-17T15:46:53.772Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:45.853 Job: Nvme0n1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:25:45.853 Verification LBA range: start 0x0 length 0x4000 00:25:45.853 Nvme0n1 : 26.65 12102.14 47.27 0.00 0.00 10557.92 168.96 3019898.88 00:25:45.853 [2024-10-17T15:46:53.772Z] =================================================================================================================== 00:25:45.853 [2024-10-17T15:46:53.772Z] Total : 12102.14 47.27 0.00 0.00 10557.92 168.96 3019898.88 00:25:45.853 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@143 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@145 -- # trap - SIGINT SIGTERM EXIT 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@147 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@148 -- # nvmftestfini 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@514 -- # nvmfcleanup 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@121 -- # sync 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@124 -- # set +e 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:46.114 rmmod nvme_tcp 00:25:46.114 rmmod nvme_fabrics 00:25:46.114 rmmod nvme_keyring 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@128 -- # set -e 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@129 -- # return 0 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@515 -- # '[' -n 172611 ']' 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@516 -- # killprocess 172611 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@950 -- # '[' -z 172611 ']' 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # kill -0 172611 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # uname 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 172611 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@968 -- # echo 'killing process with pid 172611' 00:25:46.114 killing process with pid 172611 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@969 -- # kill 172611 00:25:46.114 17:46:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@974 -- # wait 172611 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@297 -- # iptr 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@789 -- # iptables-save 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@789 -- # iptables-restore 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:46.114 17:46:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:48.651 00:25:48.651 real 0m41.210s 00:25:48.651 user 1m45.998s 00:25:48.651 sys 0m11.621s 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:25:48.651 ************************************ 00:25:48.651 END TEST nvmf_host_multipath_status 00:25:48.651 ************************************ 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@28 -- # run_test nvmf_discovery_remove_ifc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:25:48.651 ************************************ 00:25:48.651 START TEST nvmf_discovery_remove_ifc 00:25:48.651 ************************************ 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:25:48.651 * Looking for test storage... 00:25:48.651 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1691 -- # lcov --version 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # IFS=.-: 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # read -ra ver1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # IFS=.-: 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # read -ra ver2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@338 -- # local 'op=<' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@340 -- # ver1_l=2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@341 -- # ver2_l=1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@344 -- # case "$op" in 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@345 -- # : 1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # decimal 1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # ver1[v]=1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # decimal 2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # ver2[v]=2 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # return 0 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:25:48.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:48.651 --rc genhtml_branch_coverage=1 00:25:48.651 --rc genhtml_function_coverage=1 00:25:48.651 --rc genhtml_legend=1 00:25:48.651 --rc geninfo_all_blocks=1 00:25:48.651 --rc geninfo_unexecuted_blocks=1 00:25:48.651 00:25:48.651 ' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:25:48.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:48.651 --rc genhtml_branch_coverage=1 00:25:48.651 --rc genhtml_function_coverage=1 00:25:48.651 --rc genhtml_legend=1 00:25:48.651 --rc geninfo_all_blocks=1 00:25:48.651 --rc geninfo_unexecuted_blocks=1 00:25:48.651 00:25:48.651 ' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:25:48.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:48.651 --rc genhtml_branch_coverage=1 00:25:48.651 --rc genhtml_function_coverage=1 00:25:48.651 --rc genhtml_legend=1 00:25:48.651 --rc geninfo_all_blocks=1 00:25:48.651 --rc geninfo_unexecuted_blocks=1 00:25:48.651 00:25:48.651 ' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:25:48.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:48.651 --rc genhtml_branch_coverage=1 00:25:48.651 --rc genhtml_function_coverage=1 00:25:48.651 --rc genhtml_legend=1 00:25:48.651 --rc geninfo_all_blocks=1 00:25:48.651 --rc geninfo_unexecuted_blocks=1 00:25:48.651 00:25:48.651 ' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # uname -s 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@15 -- # shopt -s extglob 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:48.651 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@5 -- # export PATH 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@51 -- # : 0 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:48.652 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@14 -- # '[' tcp == rdma ']' 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@19 -- # discovery_port=8009 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@20 -- # discovery_nqn=nqn.2014-08.org.nvmexpress.discovery 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@23 -- # nqn=nqn.2016-06.io.spdk:cnode 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@25 -- # host_nqn=nqn.2021-12.io.spdk:test 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@26 -- # host_sock=/tmp/host.sock 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@39 -- # nvmftestinit 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@474 -- # prepare_net_devs 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@436 -- # local -g is_hw=no 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@438 -- # remove_spdk_ns 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@309 -- # xtrace_disable 00:25:48.652 17:46:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # pci_devs=() 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # net_devs=() 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # e810=() 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # local -ga e810 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # x722=() 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # local -ga x722 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # mlx=() 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # local -ga mlx 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:25:56.782 Found 0000:31:00.0 (0x8086 - 0x159b) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:25:56.782 Found 0000:31:00.1 (0x8086 - 0x159b) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:56.782 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:25:56.783 Found net devices under 0000:31:00.0: cvl_0_0 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:25:56.783 Found net devices under 0000:31:00.1: cvl_0_1 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # is_hw=yes 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:56.783 17:47:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:56.783 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:56.783 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.618 ms 00:25:56.783 00:25:56.783 --- 10.0.0.2 ping statistics --- 00:25:56.783 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:56.783 rtt min/avg/max/mdev = 0.618/0.618/0.618/0.000 ms 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:56.783 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:56.783 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.315 ms 00:25:56.783 00:25:56.783 --- 10.0.0.1 ping statistics --- 00:25:56.783 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:56.783 rtt min/avg/max/mdev = 0.315/0.315/0.315/0.000 ms 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@448 -- # return 0 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@40 -- # nvmfappstart -m 0x2 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@724 -- # xtrace_disable 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@507 -- # nvmfpid=182928 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@508 -- # waitforlisten 182928 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@831 -- # '[' -z 182928 ']' 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:56.783 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:56.783 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:56.783 [2024-10-17 17:47:04.126968] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:25:56.783 [2024-10-17 17:47:04.127032] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:56.783 [2024-10-17 17:47:04.217415] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:56.783 [2024-10-17 17:47:04.267057] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:56.783 [2024-10-17 17:47:04.267105] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:56.783 [2024-10-17 17:47:04.267114] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:56.783 [2024-10-17 17:47:04.267121] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:56.783 [2024-10-17 17:47:04.267128] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:56.783 [2024-10-17 17:47:04.267968] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:57.043 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:57.043 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # return 0 00:25:57.043 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:25:57.044 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@730 -- # xtrace_disable 00:25:57.044 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:57.304 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:57.304 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@43 -- # rpc_cmd 00:25:57.304 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:57.304 17:47:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:57.304 [2024-10-17 17:47:04.993680] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:57.304 [2024-10-17 17:47:05.001921] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:25:57.304 null0 00:25:57.304 [2024-10-17 17:47:05.033895] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@59 -- # hostpid=183083 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@60 -- # waitforlisten 183083 /tmp/host.sock 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock --wait-for-rpc -L bdev_nvme 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@831 -- # '[' -z 183083 ']' 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # local rpc_addr=/tmp/host.sock 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:25:57.304 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:57.304 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:57.304 [2024-10-17 17:47:05.112421] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:25:57.304 [2024-10-17 17:47:05.112484] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid183083 ] 00:25:57.304 [2024-10-17 17:47:05.194192] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:57.564 [2024-10-17 17:47:05.248554] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # return 0 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@62 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@65 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_set_options -e 1 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@66 -- # rpc_cmd -s /tmp/host.sock framework_start_init 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:58.133 17:47:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:58.133 17:47:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:58.133 17:47:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@69 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test --ctrlr-loss-timeout-sec 2 --reconnect-delay-sec 1 --fast-io-fail-timeout-sec 1 --wait-for-attach 00:25:58.133 17:47:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:58.133 17:47:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:59.511 [2024-10-17 17:47:07.082537] bdev_nvme.c:7151:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:25:59.511 [2024-10-17 17:47:07.082560] bdev_nvme.c:7237:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:25:59.511 [2024-10-17 17:47:07.082574] bdev_nvme.c:7114:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:25:59.511 [2024-10-17 17:47:07.209986] bdev_nvme.c:7080:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:25:59.511 [2024-10-17 17:47:07.273232] bdev_nvme.c:7947:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:25:59.511 [2024-10-17 17:47:07.273280] bdev_nvme.c:7947:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:25:59.511 [2024-10-17 17:47:07.273304] bdev_nvme.c:7947:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:25:59.511 [2024-10-17 17:47:07.273318] bdev_nvme.c:6970:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:25:59.512 [2024-10-17 17:47:07.273338] bdev_nvme.c:6929:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@72 -- # wait_for_bdev nvme0n1 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:59.512 [2024-10-17 17:47:07.322450] bdev_nvme.c:1736:bdev_nvme_disconnected_qpair_cb: *DEBUG*: qpair 0x113e090 was disconnected and freed. delete nvme_qpair. 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != \n\v\m\e\0\n\1 ]] 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@75 -- # ip netns exec cvl_0_0_ns_spdk ip addr del 10.0.0.2/24 dev cvl_0_0 00:25:59.512 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@76 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 down 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@79 -- # wait_for_bdev '' 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:25:59.771 17:47:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:26:00.708 17:47:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:26:02.089 17:47:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:26:03.029 17:47:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:26:03.969 17:47:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:04.909 [2024-10-17 17:47:12.713974] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 110: Connection timed out 00:26:04.909 [2024-10-17 17:47:12.714013] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:04.909 [2024-10-17 17:47:12.714022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:04.909 [2024-10-17 17:47:12.714030] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:04.909 [2024-10-17 17:47:12.714036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:04.909 [2024-10-17 17:47:12.714042] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:04.909 [2024-10-17 17:47:12.714047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:04.910 [2024-10-17 17:47:12.714053] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:04.910 [2024-10-17 17:47:12.714058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:04.910 [2024-10-17 17:47:12.714064] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:26:04.910 [2024-10-17 17:47:12.714069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:04.910 [2024-10-17 17:47:12.714075] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x111ab00 is same with the state(6) to be set 00:26:04.910 [2024-10-17 17:47:12.723995] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x111ab00 (9): Bad file descriptor 00:26:04.910 [2024-10-17 17:47:12.734031] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:26:04.910 17:47:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:04.910 17:47:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:04.910 17:47:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:04.910 17:47:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:04.910 17:47:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:04.910 17:47:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:04.910 17:47:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:06.289 [2024-10-17 17:47:13.779749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 110 00:26:06.289 [2024-10-17 17:47:13.779844] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x111ab00 with addr=10.0.0.2, port=4420 00:26:06.289 [2024-10-17 17:47:13.779889] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x111ab00 is same with the state(6) to be set 00:26:06.289 [2024-10-17 17:47:13.779945] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x111ab00 (9): Bad file descriptor 00:26:06.289 [2024-10-17 17:47:13.781053] bdev_nvme.c:3032:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:26:06.289 [2024-10-17 17:47:13.781123] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:26:06.289 [2024-10-17 17:47:13.781145] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:26:06.289 [2024-10-17 17:47:13.781170] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:26:06.289 [2024-10-17 17:47:13.781236] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:06.289 [2024-10-17 17:47:13.781261] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:26:06.289 17:47:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:06.289 17:47:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:26:06.289 17:47:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:07.232 [2024-10-17 17:47:14.783659] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:26:07.232 [2024-10-17 17:47:14.783676] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:26:07.232 [2024-10-17 17:47:14.783681] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:26:07.232 [2024-10-17 17:47:14.783686] nvme_ctrlr.c:1094:nvme_ctrlr_fail: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] already in failed state 00:26:07.232 [2024-10-17 17:47:14.783699] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:07.232 [2024-10-17 17:47:14.783713] bdev_nvme.c:6902:remove_discovery_entry: *INFO*: Discovery[10.0.0.2:8009] Remove discovery entry: nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 00:26:07.232 [2024-10-17 17:47:14.783731] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:07.232 [2024-10-17 17:47:14.783739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:07.232 [2024-10-17 17:47:14.783746] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:07.232 [2024-10-17 17:47:14.783751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:07.232 [2024-10-17 17:47:14.783757] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:07.232 [2024-10-17 17:47:14.783762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:07.232 [2024-10-17 17:47:14.783768] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:07.232 [2024-10-17 17:47:14.783774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:07.232 [2024-10-17 17:47:14.783779] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:26:07.232 [2024-10-17 17:47:14.783785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:07.232 [2024-10-17 17:47:14.783790] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery] in failed state. 00:26:07.232 [2024-10-17 17:47:14.784240] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x110a240 (9): Bad file descriptor 00:26:07.232 [2024-10-17 17:47:14.785250] nvme_fabric.c: 214:nvme_fabric_prop_get_cmd_async: *ERROR*: Failed to send Property Get fabrics command 00:26:07.232 [2024-10-17 17:47:14.785258] nvme_ctrlr.c:1213:nvme_ctrlr_shutdown_async: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery] Failed to read the CC register 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != '' ]] 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@82 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@83 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@86 -- # wait_for_bdev nvme1n1 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:07.232 17:47:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:07.232 17:47:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:26:07.232 17:47:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:08.170 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:08.170 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:08.170 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:08.170 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:08.170 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:08.170 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:08.170 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:08.171 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:08.171 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:26:08.171 17:47:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:09.110 [2024-10-17 17:47:16.797221] bdev_nvme.c:7151:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:26:09.110 [2024-10-17 17:47:16.797238] bdev_nvme.c:7237:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:26:09.110 [2024-10-17 17:47:16.797248] bdev_nvme.c:7114:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:26:09.110 [2024-10-17 17:47:16.927636] bdev_nvme.c:7080:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme1 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:26:09.370 17:47:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:26:09.370 [2024-10-17 17:47:17.152422] bdev_nvme.c:7947:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:26:09.370 [2024-10-17 17:47:17.152453] bdev_nvme.c:7947:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:26:09.370 [2024-10-17 17:47:17.152469] bdev_nvme.c:7947:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:26:09.370 [2024-10-17 17:47:17.152479] bdev_nvme.c:6970:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme1 done 00:26:09.370 [2024-10-17 17:47:17.152485] bdev_nvme.c:6929:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:26:09.370 [2024-10-17 17:47:17.196281] bdev_nvme.c:1736:bdev_nvme_disconnected_qpair_cb: *DEBUG*: qpair 0x110c4c0 was disconnected and freed. delete nvme_qpair. 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme1n1 != \n\v\m\e\1\n\1 ]] 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@88 -- # trap - SIGINT SIGTERM EXIT 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@90 -- # killprocess 183083 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@950 -- # '[' -z 183083 ']' 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # kill -0 183083 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # uname 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:10.309 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 183083 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 183083' 00:26:10.570 killing process with pid 183083 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@969 -- # kill 183083 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@974 -- # wait 183083 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@91 -- # nvmftestfini 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@121 -- # sync 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@124 -- # set +e 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:10.570 rmmod nvme_tcp 00:26:10.570 rmmod nvme_fabrics 00:26:10.570 rmmod nvme_keyring 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@128 -- # set -e 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@129 -- # return 0 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@515 -- # '[' -n 182928 ']' 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@516 -- # killprocess 182928 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@950 -- # '[' -z 182928 ']' 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # kill -0 182928 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # uname 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:10.570 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 182928 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 182928' 00:26:10.831 killing process with pid 182928 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@969 -- # kill 182928 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@974 -- # wait 182928 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@297 -- # iptr 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@789 -- # iptables-save 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@789 -- # iptables-restore 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:10.831 17:47:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:13.378 00:26:13.378 real 0m24.492s 00:26:13.378 user 0m29.430s 00:26:13.378 sys 0m7.179s 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:26:13.378 ************************************ 00:26:13.378 END TEST nvmf_discovery_remove_ifc 00:26:13.378 ************************************ 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@29 -- # run_test nvmf_identify_kernel_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:26:13.378 ************************************ 00:26:13.378 START TEST nvmf_identify_kernel_target 00:26:13.378 ************************************ 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:26:13.378 * Looking for test storage... 00:26:13.378 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1691 -- # lcov --version 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # IFS=.-: 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # read -ra ver1 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # IFS=.-: 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # read -ra ver2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@338 -- # local 'op=<' 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@340 -- # ver1_l=2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@341 -- # ver2_l=1 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@344 -- # case "$op" in 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@345 -- # : 1 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # decimal 1 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=1 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 1 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # ver1[v]=1 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # decimal 2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # ver2[v]=2 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # return 0 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:26:13.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.378 --rc genhtml_branch_coverage=1 00:26:13.378 --rc genhtml_function_coverage=1 00:26:13.378 --rc genhtml_legend=1 00:26:13.378 --rc geninfo_all_blocks=1 00:26:13.378 --rc geninfo_unexecuted_blocks=1 00:26:13.378 00:26:13.378 ' 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:26:13.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.378 --rc genhtml_branch_coverage=1 00:26:13.378 --rc genhtml_function_coverage=1 00:26:13.378 --rc genhtml_legend=1 00:26:13.378 --rc geninfo_all_blocks=1 00:26:13.378 --rc geninfo_unexecuted_blocks=1 00:26:13.378 00:26:13.378 ' 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:26:13.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.378 --rc genhtml_branch_coverage=1 00:26:13.378 --rc genhtml_function_coverage=1 00:26:13.378 --rc genhtml_legend=1 00:26:13.378 --rc geninfo_all_blocks=1 00:26:13.378 --rc geninfo_unexecuted_blocks=1 00:26:13.378 00:26:13.378 ' 00:26:13.378 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:26:13.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.378 --rc genhtml_branch_coverage=1 00:26:13.378 --rc genhtml_function_coverage=1 00:26:13.378 --rc genhtml_legend=1 00:26:13.378 --rc geninfo_all_blocks=1 00:26:13.379 --rc geninfo_unexecuted_blocks=1 00:26:13.379 00:26:13.379 ' 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # uname -s 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@15 -- # shopt -s extglob 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@5 -- # export PATH 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@51 -- # : 0 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:26:13.379 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@11 -- # nvmftestinit 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:13.379 17:47:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:13.379 17:47:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:26:13.379 17:47:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:26:13.379 17:47:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@309 -- # xtrace_disable 00:26:13.379 17:47:21 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # pci_devs=() 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # net_devs=() 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # e810=() 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # local -ga e810 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # x722=() 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # local -ga x722 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # mlx=() 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # local -ga mlx 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:21.520 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:26:21.521 Found 0000:31:00.0 (0x8086 - 0x159b) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:26:21.521 Found 0000:31:00.1 (0x8086 - 0x159b) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:26:21.521 Found net devices under 0000:31:00.0: cvl_0_0 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:26:21.521 Found net devices under 0000:31:00.1: cvl_0_1 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # is_hw=yes 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:21.521 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:21.521 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.482 ms 00:26:21.521 00:26:21.521 --- 10.0.0.2 ping statistics --- 00:26:21.521 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:21.521 rtt min/avg/max/mdev = 0.482/0.482/0.482/0.000 ms 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:21.521 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:21.521 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.301 ms 00:26:21.521 00:26:21.521 --- 10.0.0.1 ping statistics --- 00:26:21.521 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:21.521 rtt min/avg/max/mdev = 0.301/0.301/0.301/0.000 ms 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@448 -- # return 0 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@13 -- # trap 'nvmftestfini || :; clean_kernel_target' EXIT 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # get_main_ns_ip 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@767 -- # local ip 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:21.521 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # target_ip=10.0.0.1 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@16 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@658 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@660 -- # nvmet=/sys/kernel/config/nvmet 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@661 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@662 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@663 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@665 -- # local block nvme 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@667 -- # [[ ! -e /sys/module/nvmet ]] 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@668 -- # modprobe nvmet 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@671 -- # [[ -e /sys/kernel/config/nvmet ]] 00:26:21.522 17:47:28 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@673 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:26:24.162 Waiting for block devices as requested 00:26:24.466 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:26:24.466 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:26:24.466 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:26:24.466 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:26:24.726 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:26:24.726 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:26:24.726 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:26:24.986 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:26:24.986 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:26:25.246 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:26:25.246 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:26:25.246 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:26:25.507 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:26:25.507 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:26:25.507 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:26:25.768 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:26:25.768 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:26:26.028 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@676 -- # for block in /sys/block/nvme* 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@677 -- # [[ -e /sys/block/nvme0n1 ]] 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@678 -- # is_block_zoned nvme0n1 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@679 -- # block_in_use nvme0n1 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:26:26.029 No valid GPT data, bailing 00:26:26.029 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # pt= 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@395 -- # return 1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@679 -- # nvme=/dev/nvme0n1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@682 -- # [[ -b /dev/nvme0n1 ]] 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@684 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@685 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@691 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@693 -- # echo 1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@694 -- # echo /dev/nvme0n1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@695 -- # echo 1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@697 -- # echo 10.0.0.1 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@698 -- # echo tcp 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@699 -- # echo 4420 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@700 -- # echo ipv4 00:26:26.290 17:47:33 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@703 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:26:26.290 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@706 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -a 10.0.0.1 -t tcp -s 4420 00:26:26.290 00:26:26.290 Discovery Log Number of Records 2, Generation counter 2 00:26:26.290 =====Discovery Log Entry 0====== 00:26:26.290 trtype: tcp 00:26:26.290 adrfam: ipv4 00:26:26.290 subtype: current discovery subsystem 00:26:26.290 treq: not specified, sq flow control disable supported 00:26:26.290 portid: 1 00:26:26.290 trsvcid: 4420 00:26:26.290 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:26:26.290 traddr: 10.0.0.1 00:26:26.290 eflags: none 00:26:26.290 sectype: none 00:26:26.290 =====Discovery Log Entry 1====== 00:26:26.290 trtype: tcp 00:26:26.290 adrfam: ipv4 00:26:26.290 subtype: nvme subsystem 00:26:26.290 treq: not specified, sq flow control disable supported 00:26:26.290 portid: 1 00:26:26.290 trsvcid: 4420 00:26:26.290 subnqn: nqn.2016-06.io.spdk:testnqn 00:26:26.290 traddr: 10.0.0.1 00:26:26.290 eflags: none 00:26:26.290 sectype: none 00:26:26.290 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 00:26:26.290 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' 00:26:26.290 ===================================================== 00:26:26.290 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2014-08.org.nvmexpress.discovery 00:26:26.290 ===================================================== 00:26:26.290 Controller Capabilities/Features 00:26:26.290 ================================ 00:26:26.290 Vendor ID: 0000 00:26:26.290 Subsystem Vendor ID: 0000 00:26:26.290 Serial Number: bb6a0cd1dd4dfcb64385 00:26:26.290 Model Number: Linux 00:26:26.290 Firmware Version: 6.8.9-20 00:26:26.290 Recommended Arb Burst: 0 00:26:26.290 IEEE OUI Identifier: 00 00 00 00:26:26.290 Multi-path I/O 00:26:26.290 May have multiple subsystem ports: No 00:26:26.290 May have multiple controllers: No 00:26:26.290 Associated with SR-IOV VF: No 00:26:26.290 Max Data Transfer Size: Unlimited 00:26:26.290 Max Number of Namespaces: 0 00:26:26.290 Max Number of I/O Queues: 1024 00:26:26.290 NVMe Specification Version (VS): 1.3 00:26:26.290 NVMe Specification Version (Identify): 1.3 00:26:26.290 Maximum Queue Entries: 1024 00:26:26.290 Contiguous Queues Required: No 00:26:26.290 Arbitration Mechanisms Supported 00:26:26.290 Weighted Round Robin: Not Supported 00:26:26.290 Vendor Specific: Not Supported 00:26:26.290 Reset Timeout: 7500 ms 00:26:26.290 Doorbell Stride: 4 bytes 00:26:26.290 NVM Subsystem Reset: Not Supported 00:26:26.290 Command Sets Supported 00:26:26.290 NVM Command Set: Supported 00:26:26.290 Boot Partition: Not Supported 00:26:26.290 Memory Page Size Minimum: 4096 bytes 00:26:26.290 Memory Page Size Maximum: 4096 bytes 00:26:26.290 Persistent Memory Region: Not Supported 00:26:26.290 Optional Asynchronous Events Supported 00:26:26.290 Namespace Attribute Notices: Not Supported 00:26:26.290 Firmware Activation Notices: Not Supported 00:26:26.290 ANA Change Notices: Not Supported 00:26:26.290 PLE Aggregate Log Change Notices: Not Supported 00:26:26.290 LBA Status Info Alert Notices: Not Supported 00:26:26.290 EGE Aggregate Log Change Notices: Not Supported 00:26:26.290 Normal NVM Subsystem Shutdown event: Not Supported 00:26:26.290 Zone Descriptor Change Notices: Not Supported 00:26:26.290 Discovery Log Change Notices: Supported 00:26:26.290 Controller Attributes 00:26:26.290 128-bit Host Identifier: Not Supported 00:26:26.290 Non-Operational Permissive Mode: Not Supported 00:26:26.290 NVM Sets: Not Supported 00:26:26.290 Read Recovery Levels: Not Supported 00:26:26.290 Endurance Groups: Not Supported 00:26:26.290 Predictable Latency Mode: Not Supported 00:26:26.290 Traffic Based Keep ALive: Not Supported 00:26:26.290 Namespace Granularity: Not Supported 00:26:26.290 SQ Associations: Not Supported 00:26:26.290 UUID List: Not Supported 00:26:26.290 Multi-Domain Subsystem: Not Supported 00:26:26.290 Fixed Capacity Management: Not Supported 00:26:26.290 Variable Capacity Management: Not Supported 00:26:26.290 Delete Endurance Group: Not Supported 00:26:26.290 Delete NVM Set: Not Supported 00:26:26.290 Extended LBA Formats Supported: Not Supported 00:26:26.290 Flexible Data Placement Supported: Not Supported 00:26:26.290 00:26:26.290 Controller Memory Buffer Support 00:26:26.290 ================================ 00:26:26.290 Supported: No 00:26:26.290 00:26:26.290 Persistent Memory Region Support 00:26:26.290 ================================ 00:26:26.290 Supported: No 00:26:26.290 00:26:26.290 Admin Command Set Attributes 00:26:26.291 ============================ 00:26:26.291 Security Send/Receive: Not Supported 00:26:26.291 Format NVM: Not Supported 00:26:26.291 Firmware Activate/Download: Not Supported 00:26:26.291 Namespace Management: Not Supported 00:26:26.291 Device Self-Test: Not Supported 00:26:26.291 Directives: Not Supported 00:26:26.291 NVMe-MI: Not Supported 00:26:26.291 Virtualization Management: Not Supported 00:26:26.291 Doorbell Buffer Config: Not Supported 00:26:26.291 Get LBA Status Capability: Not Supported 00:26:26.291 Command & Feature Lockdown Capability: Not Supported 00:26:26.291 Abort Command Limit: 1 00:26:26.291 Async Event Request Limit: 1 00:26:26.291 Number of Firmware Slots: N/A 00:26:26.291 Firmware Slot 1 Read-Only: N/A 00:26:26.291 Firmware Activation Without Reset: N/A 00:26:26.291 Multiple Update Detection Support: N/A 00:26:26.291 Firmware Update Granularity: No Information Provided 00:26:26.291 Per-Namespace SMART Log: No 00:26:26.291 Asymmetric Namespace Access Log Page: Not Supported 00:26:26.291 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:26:26.291 Command Effects Log Page: Not Supported 00:26:26.291 Get Log Page Extended Data: Supported 00:26:26.291 Telemetry Log Pages: Not Supported 00:26:26.291 Persistent Event Log Pages: Not Supported 00:26:26.291 Supported Log Pages Log Page: May Support 00:26:26.291 Commands Supported & Effects Log Page: Not Supported 00:26:26.291 Feature Identifiers & Effects Log Page:May Support 00:26:26.291 NVMe-MI Commands & Effects Log Page: May Support 00:26:26.291 Data Area 4 for Telemetry Log: Not Supported 00:26:26.291 Error Log Page Entries Supported: 1 00:26:26.291 Keep Alive: Not Supported 00:26:26.291 00:26:26.291 NVM Command Set Attributes 00:26:26.291 ========================== 00:26:26.291 Submission Queue Entry Size 00:26:26.291 Max: 1 00:26:26.291 Min: 1 00:26:26.291 Completion Queue Entry Size 00:26:26.291 Max: 1 00:26:26.291 Min: 1 00:26:26.291 Number of Namespaces: 0 00:26:26.291 Compare Command: Not Supported 00:26:26.291 Write Uncorrectable Command: Not Supported 00:26:26.291 Dataset Management Command: Not Supported 00:26:26.291 Write Zeroes Command: Not Supported 00:26:26.291 Set Features Save Field: Not Supported 00:26:26.291 Reservations: Not Supported 00:26:26.291 Timestamp: Not Supported 00:26:26.291 Copy: Not Supported 00:26:26.291 Volatile Write Cache: Not Present 00:26:26.291 Atomic Write Unit (Normal): 1 00:26:26.291 Atomic Write Unit (PFail): 1 00:26:26.291 Atomic Compare & Write Unit: 1 00:26:26.291 Fused Compare & Write: Not Supported 00:26:26.291 Scatter-Gather List 00:26:26.291 SGL Command Set: Supported 00:26:26.291 SGL Keyed: Not Supported 00:26:26.291 SGL Bit Bucket Descriptor: Not Supported 00:26:26.291 SGL Metadata Pointer: Not Supported 00:26:26.291 Oversized SGL: Not Supported 00:26:26.291 SGL Metadata Address: Not Supported 00:26:26.291 SGL Offset: Supported 00:26:26.291 Transport SGL Data Block: Not Supported 00:26:26.291 Replay Protected Memory Block: Not Supported 00:26:26.291 00:26:26.291 Firmware Slot Information 00:26:26.291 ========================= 00:26:26.291 Active slot: 0 00:26:26.291 00:26:26.291 00:26:26.291 Error Log 00:26:26.291 ========= 00:26:26.291 00:26:26.291 Active Namespaces 00:26:26.291 ================= 00:26:26.291 Discovery Log Page 00:26:26.291 ================== 00:26:26.291 Generation Counter: 2 00:26:26.291 Number of Records: 2 00:26:26.291 Record Format: 0 00:26:26.291 00:26:26.291 Discovery Log Entry 0 00:26:26.291 ---------------------- 00:26:26.291 Transport Type: 3 (TCP) 00:26:26.291 Address Family: 1 (IPv4) 00:26:26.291 Subsystem Type: 3 (Current Discovery Subsystem) 00:26:26.291 Entry Flags: 00:26:26.291 Duplicate Returned Information: 0 00:26:26.291 Explicit Persistent Connection Support for Discovery: 0 00:26:26.291 Transport Requirements: 00:26:26.291 Secure Channel: Not Specified 00:26:26.291 Port ID: 1 (0x0001) 00:26:26.291 Controller ID: 65535 (0xffff) 00:26:26.291 Admin Max SQ Size: 32 00:26:26.291 Transport Service Identifier: 4420 00:26:26.291 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:26:26.291 Transport Address: 10.0.0.1 00:26:26.291 Discovery Log Entry 1 00:26:26.291 ---------------------- 00:26:26.291 Transport Type: 3 (TCP) 00:26:26.291 Address Family: 1 (IPv4) 00:26:26.291 Subsystem Type: 2 (NVM Subsystem) 00:26:26.291 Entry Flags: 00:26:26.291 Duplicate Returned Information: 0 00:26:26.291 Explicit Persistent Connection Support for Discovery: 0 00:26:26.291 Transport Requirements: 00:26:26.291 Secure Channel: Not Specified 00:26:26.291 Port ID: 1 (0x0001) 00:26:26.291 Controller ID: 65535 (0xffff) 00:26:26.291 Admin Max SQ Size: 32 00:26:26.291 Transport Service Identifier: 4420 00:26:26.291 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:testnqn 00:26:26.291 Transport Address: 10.0.0.1 00:26:26.291 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:26:26.553 get_feature(0x01) failed 00:26:26.553 get_feature(0x02) failed 00:26:26.553 get_feature(0x04) failed 00:26:26.553 ===================================================== 00:26:26.553 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:26:26.553 ===================================================== 00:26:26.553 Controller Capabilities/Features 00:26:26.553 ================================ 00:26:26.553 Vendor ID: 0000 00:26:26.553 Subsystem Vendor ID: 0000 00:26:26.553 Serial Number: 0ddc6c06da0be454743a 00:26:26.553 Model Number: SPDK-nqn.2016-06.io.spdk:testnqn 00:26:26.553 Firmware Version: 6.8.9-20 00:26:26.553 Recommended Arb Burst: 6 00:26:26.553 IEEE OUI Identifier: 00 00 00 00:26:26.553 Multi-path I/O 00:26:26.553 May have multiple subsystem ports: Yes 00:26:26.553 May have multiple controllers: Yes 00:26:26.553 Associated with SR-IOV VF: No 00:26:26.553 Max Data Transfer Size: Unlimited 00:26:26.553 Max Number of Namespaces: 1024 00:26:26.553 Max Number of I/O Queues: 128 00:26:26.553 NVMe Specification Version (VS): 1.3 00:26:26.553 NVMe Specification Version (Identify): 1.3 00:26:26.553 Maximum Queue Entries: 1024 00:26:26.553 Contiguous Queues Required: No 00:26:26.553 Arbitration Mechanisms Supported 00:26:26.553 Weighted Round Robin: Not Supported 00:26:26.553 Vendor Specific: Not Supported 00:26:26.553 Reset Timeout: 7500 ms 00:26:26.553 Doorbell Stride: 4 bytes 00:26:26.553 NVM Subsystem Reset: Not Supported 00:26:26.553 Command Sets Supported 00:26:26.553 NVM Command Set: Supported 00:26:26.553 Boot Partition: Not Supported 00:26:26.553 Memory Page Size Minimum: 4096 bytes 00:26:26.553 Memory Page Size Maximum: 4096 bytes 00:26:26.553 Persistent Memory Region: Not Supported 00:26:26.553 Optional Asynchronous Events Supported 00:26:26.553 Namespace Attribute Notices: Supported 00:26:26.553 Firmware Activation Notices: Not Supported 00:26:26.553 ANA Change Notices: Supported 00:26:26.553 PLE Aggregate Log Change Notices: Not Supported 00:26:26.553 LBA Status Info Alert Notices: Not Supported 00:26:26.553 EGE Aggregate Log Change Notices: Not Supported 00:26:26.553 Normal NVM Subsystem Shutdown event: Not Supported 00:26:26.553 Zone Descriptor Change Notices: Not Supported 00:26:26.553 Discovery Log Change Notices: Not Supported 00:26:26.553 Controller Attributes 00:26:26.553 128-bit Host Identifier: Supported 00:26:26.553 Non-Operational Permissive Mode: Not Supported 00:26:26.553 NVM Sets: Not Supported 00:26:26.553 Read Recovery Levels: Not Supported 00:26:26.553 Endurance Groups: Not Supported 00:26:26.553 Predictable Latency Mode: Not Supported 00:26:26.553 Traffic Based Keep ALive: Supported 00:26:26.553 Namespace Granularity: Not Supported 00:26:26.553 SQ Associations: Not Supported 00:26:26.553 UUID List: Not Supported 00:26:26.553 Multi-Domain Subsystem: Not Supported 00:26:26.553 Fixed Capacity Management: Not Supported 00:26:26.553 Variable Capacity Management: Not Supported 00:26:26.553 Delete Endurance Group: Not Supported 00:26:26.553 Delete NVM Set: Not Supported 00:26:26.553 Extended LBA Formats Supported: Not Supported 00:26:26.553 Flexible Data Placement Supported: Not Supported 00:26:26.553 00:26:26.553 Controller Memory Buffer Support 00:26:26.553 ================================ 00:26:26.553 Supported: No 00:26:26.554 00:26:26.554 Persistent Memory Region Support 00:26:26.554 ================================ 00:26:26.554 Supported: No 00:26:26.554 00:26:26.554 Admin Command Set Attributes 00:26:26.554 ============================ 00:26:26.554 Security Send/Receive: Not Supported 00:26:26.554 Format NVM: Not Supported 00:26:26.554 Firmware Activate/Download: Not Supported 00:26:26.554 Namespace Management: Not Supported 00:26:26.554 Device Self-Test: Not Supported 00:26:26.554 Directives: Not Supported 00:26:26.554 NVMe-MI: Not Supported 00:26:26.554 Virtualization Management: Not Supported 00:26:26.554 Doorbell Buffer Config: Not Supported 00:26:26.554 Get LBA Status Capability: Not Supported 00:26:26.554 Command & Feature Lockdown Capability: Not Supported 00:26:26.554 Abort Command Limit: 4 00:26:26.554 Async Event Request Limit: 4 00:26:26.554 Number of Firmware Slots: N/A 00:26:26.554 Firmware Slot 1 Read-Only: N/A 00:26:26.554 Firmware Activation Without Reset: N/A 00:26:26.554 Multiple Update Detection Support: N/A 00:26:26.554 Firmware Update Granularity: No Information Provided 00:26:26.554 Per-Namespace SMART Log: Yes 00:26:26.554 Asymmetric Namespace Access Log Page: Supported 00:26:26.554 ANA Transition Time : 10 sec 00:26:26.554 00:26:26.554 Asymmetric Namespace Access Capabilities 00:26:26.554 ANA Optimized State : Supported 00:26:26.554 ANA Non-Optimized State : Supported 00:26:26.554 ANA Inaccessible State : Supported 00:26:26.554 ANA Persistent Loss State : Supported 00:26:26.554 ANA Change State : Supported 00:26:26.554 ANAGRPID is not changed : No 00:26:26.554 Non-Zero ANAGRPID for NS Mgmt Cmd : Not Supported 00:26:26.554 00:26:26.554 ANA Group Identifier Maximum : 128 00:26:26.554 Number of ANA Group Identifiers : 128 00:26:26.554 Max Number of Allowed Namespaces : 1024 00:26:26.554 Subsystem NQN: nqn.2016-06.io.spdk:testnqn 00:26:26.554 Command Effects Log Page: Supported 00:26:26.554 Get Log Page Extended Data: Supported 00:26:26.554 Telemetry Log Pages: Not Supported 00:26:26.554 Persistent Event Log Pages: Not Supported 00:26:26.554 Supported Log Pages Log Page: May Support 00:26:26.554 Commands Supported & Effects Log Page: Not Supported 00:26:26.554 Feature Identifiers & Effects Log Page:May Support 00:26:26.554 NVMe-MI Commands & Effects Log Page: May Support 00:26:26.554 Data Area 4 for Telemetry Log: Not Supported 00:26:26.554 Error Log Page Entries Supported: 128 00:26:26.554 Keep Alive: Supported 00:26:26.554 Keep Alive Granularity: 1000 ms 00:26:26.554 00:26:26.554 NVM Command Set Attributes 00:26:26.554 ========================== 00:26:26.554 Submission Queue Entry Size 00:26:26.554 Max: 64 00:26:26.554 Min: 64 00:26:26.554 Completion Queue Entry Size 00:26:26.554 Max: 16 00:26:26.554 Min: 16 00:26:26.554 Number of Namespaces: 1024 00:26:26.554 Compare Command: Not Supported 00:26:26.554 Write Uncorrectable Command: Not Supported 00:26:26.554 Dataset Management Command: Supported 00:26:26.554 Write Zeroes Command: Supported 00:26:26.554 Set Features Save Field: Not Supported 00:26:26.554 Reservations: Not Supported 00:26:26.554 Timestamp: Not Supported 00:26:26.554 Copy: Not Supported 00:26:26.554 Volatile Write Cache: Present 00:26:26.554 Atomic Write Unit (Normal): 1 00:26:26.554 Atomic Write Unit (PFail): 1 00:26:26.554 Atomic Compare & Write Unit: 1 00:26:26.554 Fused Compare & Write: Not Supported 00:26:26.554 Scatter-Gather List 00:26:26.554 SGL Command Set: Supported 00:26:26.554 SGL Keyed: Not Supported 00:26:26.554 SGL Bit Bucket Descriptor: Not Supported 00:26:26.554 SGL Metadata Pointer: Not Supported 00:26:26.554 Oversized SGL: Not Supported 00:26:26.554 SGL Metadata Address: Not Supported 00:26:26.554 SGL Offset: Supported 00:26:26.554 Transport SGL Data Block: Not Supported 00:26:26.554 Replay Protected Memory Block: Not Supported 00:26:26.554 00:26:26.554 Firmware Slot Information 00:26:26.554 ========================= 00:26:26.554 Active slot: 0 00:26:26.554 00:26:26.554 Asymmetric Namespace Access 00:26:26.554 =========================== 00:26:26.554 Change Count : 0 00:26:26.554 Number of ANA Group Descriptors : 1 00:26:26.554 ANA Group Descriptor : 0 00:26:26.554 ANA Group ID : 1 00:26:26.554 Number of NSID Values : 1 00:26:26.554 Change Count : 0 00:26:26.554 ANA State : 1 00:26:26.554 Namespace Identifier : 1 00:26:26.554 00:26:26.554 Commands Supported and Effects 00:26:26.554 ============================== 00:26:26.554 Admin Commands 00:26:26.554 -------------- 00:26:26.554 Get Log Page (02h): Supported 00:26:26.554 Identify (06h): Supported 00:26:26.554 Abort (08h): Supported 00:26:26.554 Set Features (09h): Supported 00:26:26.554 Get Features (0Ah): Supported 00:26:26.554 Asynchronous Event Request (0Ch): Supported 00:26:26.554 Keep Alive (18h): Supported 00:26:26.554 I/O Commands 00:26:26.554 ------------ 00:26:26.554 Flush (00h): Supported 00:26:26.554 Write (01h): Supported LBA-Change 00:26:26.554 Read (02h): Supported 00:26:26.554 Write Zeroes (08h): Supported LBA-Change 00:26:26.554 Dataset Management (09h): Supported 00:26:26.554 00:26:26.554 Error Log 00:26:26.554 ========= 00:26:26.554 Entry: 0 00:26:26.554 Error Count: 0x3 00:26:26.554 Submission Queue Id: 0x0 00:26:26.554 Command Id: 0x5 00:26:26.554 Phase Bit: 0 00:26:26.554 Status Code: 0x2 00:26:26.554 Status Code Type: 0x0 00:26:26.554 Do Not Retry: 1 00:26:26.554 Error Location: 0x28 00:26:26.554 LBA: 0x0 00:26:26.554 Namespace: 0x0 00:26:26.554 Vendor Log Page: 0x0 00:26:26.554 ----------- 00:26:26.554 Entry: 1 00:26:26.554 Error Count: 0x2 00:26:26.554 Submission Queue Id: 0x0 00:26:26.554 Command Id: 0x5 00:26:26.554 Phase Bit: 0 00:26:26.554 Status Code: 0x2 00:26:26.554 Status Code Type: 0x0 00:26:26.554 Do Not Retry: 1 00:26:26.554 Error Location: 0x28 00:26:26.554 LBA: 0x0 00:26:26.554 Namespace: 0x0 00:26:26.554 Vendor Log Page: 0x0 00:26:26.554 ----------- 00:26:26.554 Entry: 2 00:26:26.554 Error Count: 0x1 00:26:26.554 Submission Queue Id: 0x0 00:26:26.554 Command Id: 0x4 00:26:26.554 Phase Bit: 0 00:26:26.554 Status Code: 0x2 00:26:26.554 Status Code Type: 0x0 00:26:26.554 Do Not Retry: 1 00:26:26.554 Error Location: 0x28 00:26:26.554 LBA: 0x0 00:26:26.554 Namespace: 0x0 00:26:26.554 Vendor Log Page: 0x0 00:26:26.554 00:26:26.554 Number of Queues 00:26:26.554 ================ 00:26:26.554 Number of I/O Submission Queues: 128 00:26:26.554 Number of I/O Completion Queues: 128 00:26:26.554 00:26:26.554 ZNS Specific Controller Data 00:26:26.554 ============================ 00:26:26.554 Zone Append Size Limit: 0 00:26:26.554 00:26:26.554 00:26:26.554 Active Namespaces 00:26:26.554 ================= 00:26:26.554 get_feature(0x05) failed 00:26:26.554 Namespace ID:1 00:26:26.554 Command Set Identifier: NVM (00h) 00:26:26.554 Deallocate: Supported 00:26:26.554 Deallocated/Unwritten Error: Not Supported 00:26:26.554 Deallocated Read Value: Unknown 00:26:26.554 Deallocate in Write Zeroes: Not Supported 00:26:26.554 Deallocated Guard Field: 0xFFFF 00:26:26.554 Flush: Supported 00:26:26.554 Reservation: Not Supported 00:26:26.554 Namespace Sharing Capabilities: Multiple Controllers 00:26:26.554 Size (in LBAs): 3750748848 (1788GiB) 00:26:26.554 Capacity (in LBAs): 3750748848 (1788GiB) 00:26:26.554 Utilization (in LBAs): 3750748848 (1788GiB) 00:26:26.554 UUID: 9487238d-33ec-45fe-8919-310fceaaa572 00:26:26.554 Thin Provisioning: Not Supported 00:26:26.554 Per-NS Atomic Units: Yes 00:26:26.554 Atomic Write Unit (Normal): 8 00:26:26.554 Atomic Write Unit (PFail): 8 00:26:26.554 Preferred Write Granularity: 8 00:26:26.554 Atomic Compare & Write Unit: 8 00:26:26.554 Atomic Boundary Size (Normal): 0 00:26:26.554 Atomic Boundary Size (PFail): 0 00:26:26.554 Atomic Boundary Offset: 0 00:26:26.554 NGUID/EUI64 Never Reused: No 00:26:26.554 ANA group ID: 1 00:26:26.554 Namespace Write Protected: No 00:26:26.554 Number of LBA Formats: 1 00:26:26.554 Current LBA Format: LBA Format #00 00:26:26.555 LBA Format #00: Data Size: 512 Metadata Size: 0 00:26:26.555 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # nvmftestfini 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@121 -- # sync 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@124 -- # set +e 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:26.555 rmmod nvme_tcp 00:26:26.555 rmmod nvme_fabrics 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@128 -- # set -e 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@129 -- # return 0 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@297 -- # iptr 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@789 -- # iptables-save 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@789 -- # iptables-restore 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:26.555 17:47:34 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # clean_kernel_target 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@710 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@712 -- # echo 0 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@714 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@715 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@716 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@719 -- # modules=(/sys/module/nvmet/holders/*) 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@721 -- # modprobe -r nvmet_tcp nvmet 00:26:29.103 17:47:36 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@724 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:26:32.407 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:26:32.407 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:26:32.979 00:26:32.979 real 0m19.871s 00:26:32.979 user 0m5.392s 00:26:32.979 sys 0m11.417s 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:26:32.979 ************************************ 00:26:32.979 END TEST nvmf_identify_kernel_target 00:26:32.979 ************************************ 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@30 -- # run_test nvmf_auth_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:26:32.979 ************************************ 00:26:32.979 START TEST nvmf_auth_host 00:26:32.979 ************************************ 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:26:32.979 * Looking for test storage... 00:26:32.979 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1691 -- # lcov --version 00:26:32.979 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # IFS=.-: 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # read -ra ver1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # IFS=.-: 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # read -ra ver2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@338 -- # local 'op=<' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@340 -- # ver1_l=2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@341 -- # ver2_l=1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@344 -- # case "$op" in 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@345 -- # : 1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # decimal 1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # ver1[v]=1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # decimal 2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # ver2[v]=2 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # return 0 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:26:33.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:33.241 --rc genhtml_branch_coverage=1 00:26:33.241 --rc genhtml_function_coverage=1 00:26:33.241 --rc genhtml_legend=1 00:26:33.241 --rc geninfo_all_blocks=1 00:26:33.241 --rc geninfo_unexecuted_blocks=1 00:26:33.241 00:26:33.241 ' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:26:33.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:33.241 --rc genhtml_branch_coverage=1 00:26:33.241 --rc genhtml_function_coverage=1 00:26:33.241 --rc genhtml_legend=1 00:26:33.241 --rc geninfo_all_blocks=1 00:26:33.241 --rc geninfo_unexecuted_blocks=1 00:26:33.241 00:26:33.241 ' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:26:33.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:33.241 --rc genhtml_branch_coverage=1 00:26:33.241 --rc genhtml_function_coverage=1 00:26:33.241 --rc genhtml_legend=1 00:26:33.241 --rc geninfo_all_blocks=1 00:26:33.241 --rc geninfo_unexecuted_blocks=1 00:26:33.241 00:26:33.241 ' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:26:33.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:33.241 --rc genhtml_branch_coverage=1 00:26:33.241 --rc genhtml_function_coverage=1 00:26:33.241 --rc genhtml_legend=1 00:26:33.241 --rc geninfo_all_blocks=1 00:26:33.241 --rc geninfo_unexecuted_blocks=1 00:26:33.241 00:26:33.241 ' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # uname -s 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@15 -- # shopt -s extglob 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:33.241 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@5 -- # export PATH 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@51 -- # : 0 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:26:33.242 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@16 -- # dhgroups=("ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@17 -- # subnqn=nqn.2024-02.io.spdk:cnode0 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@18 -- # hostnqn=nqn.2024-02.io.spdk:host0 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@19 -- # nvmet_subsys=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@20 -- # nvmet_host=/sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # keys=() 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # ckeys=() 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@68 -- # nvmftestinit 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@474 -- # prepare_net_devs 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@436 -- # local -g is_hw=no 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@438 -- # remove_spdk_ns 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@309 -- # xtrace_disable 00:26:33.242 17:47:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # pci_devs=() 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # net_devs=() 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # e810=() 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # local -ga e810 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # x722=() 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # local -ga x722 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # mlx=() 00:26:41.387 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # local -ga mlx 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:26:41.388 Found 0000:31:00.0 (0x8086 - 0x159b) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:26:41.388 Found 0000:31:00.1 (0x8086 - 0x159b) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:26:41.388 Found net devices under 0000:31:00.0: cvl_0_0 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:26:41.388 Found net devices under 0000:31:00.1: cvl_0_1 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # is_hw=yes 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:41.388 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:41.388 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.658 ms 00:26:41.388 00:26:41.388 --- 10.0.0.2 ping statistics --- 00:26:41.388 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:41.388 rtt min/avg/max/mdev = 0.658/0.658/0.658/0.000 ms 00:26:41.388 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:41.388 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:41.388 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.320 ms 00:26:41.388 00:26:41.388 --- 10.0.0.1 ping statistics --- 00:26:41.389 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:41.389 rtt min/avg/max/mdev = 0.320/0.320/0.320/0.000 ms 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@448 -- # return 0 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@69 -- # nvmfappstart -L nvme_auth 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@507 -- # nvmfpid=197856 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@508 -- # waitforlisten 197856 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvme_auth 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@831 -- # '[' -z 197856 ']' 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:41.389 17:47:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # return 0 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@70 -- # trap 'cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log; cleanup' SIGINT SIGTERM EXIT 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key null 32 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=null 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=e7fd44a3d8c82e9b10235c99a958d4aa 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.9du 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key e7fd44a3d8c82e9b10235c99a958d4aa 0 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 e7fd44a3d8c82e9b10235c99a958d4aa 0 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:41.649 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:41.650 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=e7fd44a3d8c82e9b10235c99a958d4aa 00:26:41.650 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=0 00:26:41.650 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.9du 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.9du 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # keys[0]=/tmp/spdk.key-null.9du 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key sha512 64 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha512 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=64 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=3d9da57fb273ffcdd530a87ce1ed4301e0fd9d24cca610e876529ade1a8b7163 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.lTw 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 3d9da57fb273ffcdd530a87ce1ed4301e0fd9d24cca610e876529ade1a8b7163 3 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 3d9da57fb273ffcdd530a87ce1ed4301e0fd9d24cca610e876529ade1a8b7163 3 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=3d9da57fb273ffcdd530a87ce1ed4301e0fd9d24cca610e876529ade1a8b7163 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=3 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.lTw 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.lTw 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # ckeys[0]=/tmp/spdk.key-sha512.lTw 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key null 48 00:26:41.911 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=null 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=48 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=acef2c7336ef878ba1639566d1fbcf8a6cc342cfa50e4cf0 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.uDJ 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key acef2c7336ef878ba1639566d1fbcf8a6cc342cfa50e4cf0 0 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 acef2c7336ef878ba1639566d1fbcf8a6cc342cfa50e4cf0 0 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=acef2c7336ef878ba1639566d1fbcf8a6cc342cfa50e4cf0 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=0 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.uDJ 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.uDJ 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # keys[1]=/tmp/spdk.key-null.uDJ 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key sha384 48 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha384 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=48 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=cc190e6a181e5b0db7774ab46b9fe02c1adbd2c1cc65e6de 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.zQR 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key cc190e6a181e5b0db7774ab46b9fe02c1adbd2c1cc65e6de 2 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 cc190e6a181e5b0db7774ab46b9fe02c1adbd2c1cc65e6de 2 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=cc190e6a181e5b0db7774ab46b9fe02c1adbd2c1cc65e6de 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=2 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.zQR 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.zQR 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # ckeys[1]=/tmp/spdk.key-sha384.zQR 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha256 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:26:41.912 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=665f4126796f30fd2399948b6b495d31 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.Hy8 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 665f4126796f30fd2399948b6b495d31 1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 665f4126796f30fd2399948b6b495d31 1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=665f4126796f30fd2399948b6b495d31 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.Hy8 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.Hy8 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # keys[2]=/tmp/spdk.key-sha256.Hy8 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha256 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=39d217674256d88e2e6ce2424f4ed914 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.e09 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 39d217674256d88e2e6ce2424f4ed914 1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 39d217674256d88e2e6ce2424f4ed914 1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=39d217674256d88e2e6ce2424f4ed914 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.e09 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.e09 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # ckeys[2]=/tmp/spdk.key-sha256.e09 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key sha384 48 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha384 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=48 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=a3eaaf381117795f54ef8fa7fd6808cb715b5b83aff9d844 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.OMZ 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key a3eaaf381117795f54ef8fa7fd6808cb715b5b83aff9d844 2 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 a3eaaf381117795f54ef8fa7fd6808cb715b5b83aff9d844 2 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:42.174 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=a3eaaf381117795f54ef8fa7fd6808cb715b5b83aff9d844 00:26:42.175 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=2 00:26:42.175 17:47:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.OMZ 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.OMZ 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # keys[3]=/tmp/spdk.key-sha384.OMZ 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key null 32 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=null 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=9831a1cefc1e156f6ce75e0dccf2a76f 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.RxM 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 9831a1cefc1e156f6ce75e0dccf2a76f 0 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 9831a1cefc1e156f6ce75e0dccf2a76f 0 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=9831a1cefc1e156f6ce75e0dccf2a76f 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=0 00:26:42.175 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.RxM 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.RxM 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # ckeys[3]=/tmp/spdk.key-null.RxM 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # gen_dhchap_key sha512 64 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha512 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=64 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=13953dd83a170ed18d2a62df4bd3be4e492004aecad4d79dfe9adb9d458d41e6 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.wIx 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 13953dd83a170ed18d2a62df4bd3be4e492004aecad4d79dfe9adb9d458d41e6 3 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 13953dd83a170ed18d2a62df4bd3be4e492004aecad4d79dfe9adb9d458d41e6 3 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=13953dd83a170ed18d2a62df4bd3be4e492004aecad4d79dfe9adb9d458d41e6 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=3 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.wIx 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.wIx 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # keys[4]=/tmp/spdk.key-sha512.wIx 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # ckeys[4]= 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@79 -- # waitforlisten 197856 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@831 -- # '[' -z 197856 ']' 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:42.437 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:42.437 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # return 0 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.9du 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha512.lTw ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.lTw 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-null.uDJ 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha384.zQR ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.zQR 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha256.Hy8 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha256.e09 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.e09 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha384.OMZ 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-null.RxM ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey3 /tmp/spdk.key-null.RxM 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key4 /tmp/spdk.key-sha512.wIx 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n '' ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@85 -- # nvmet_auth_init 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # get_main_ns_ip 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # configure_kernel_target nqn.2024-02.io.spdk:cnode0 10.0.0.1 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@658 -- # local kernel_name=nqn.2024-02.io.spdk:cnode0 kernel_target_ip=10.0.0.1 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@660 -- # nvmet=/sys/kernel/config/nvmet 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@661 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@662 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@663 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@665 -- # local block nvme 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@667 -- # [[ ! -e /sys/module/nvmet ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@668 -- # modprobe nvmet 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@671 -- # [[ -e /sys/kernel/config/nvmet ]] 00:26:42.699 17:47:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@673 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:26:46.005 Waiting for block devices as requested 00:26:46.265 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:26:46.265 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:26:46.265 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:26:46.527 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:26:46.527 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:26:46.527 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:26:46.527 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:26:46.786 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:26:46.786 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:26:47.047 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:26:47.047 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:26:47.047 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:26:47.307 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:26:47.307 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:26:47.307 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:26:47.307 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:26:47.568 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@676 -- # for block in /sys/block/nvme* 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@677 -- # [[ -e /sys/block/nvme0n1 ]] 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@678 -- # is_block_zoned nvme0n1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@679 -- # block_in_use nvme0n1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:26:48.507 No valid GPT data, bailing 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # pt= 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@395 -- # return 1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@679 -- # nvme=/dev/nvme0n1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@682 -- # [[ -b /dev/nvme0n1 ]] 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@684 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@685 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@691 -- # echo SPDK-nqn.2024-02.io.spdk:cnode0 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@693 -- # echo 1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@694 -- # echo /dev/nvme0n1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@695 -- # echo 1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@697 -- # echo 10.0.0.1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@698 -- # echo tcp 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@699 -- # echo 4420 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@700 -- # echo ipv4 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@703 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 /sys/kernel/config/nvmet/ports/1/subsystems/ 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@706 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -a 10.0.0.1 -t tcp -s 4420 00:26:48.507 00:26:48.507 Discovery Log Number of Records 2, Generation counter 2 00:26:48.507 =====Discovery Log Entry 0====== 00:26:48.507 trtype: tcp 00:26:48.507 adrfam: ipv4 00:26:48.507 subtype: current discovery subsystem 00:26:48.507 treq: not specified, sq flow control disable supported 00:26:48.507 portid: 1 00:26:48.507 trsvcid: 4420 00:26:48.507 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:26:48.507 traddr: 10.0.0.1 00:26:48.507 eflags: none 00:26:48.507 sectype: none 00:26:48.507 =====Discovery Log Entry 1====== 00:26:48.507 trtype: tcp 00:26:48.507 adrfam: ipv4 00:26:48.507 subtype: nvme subsystem 00:26:48.507 treq: not specified, sq flow control disable supported 00:26:48.507 portid: 1 00:26:48.507 trsvcid: 4420 00:26:48.507 subnqn: nqn.2024-02.io.spdk:cnode0 00:26:48.507 traddr: 10.0.0.1 00:26:48.507 eflags: none 00:26:48.507 sectype: none 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@36 -- # mkdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@37 -- # echo 0 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@38 -- # ln -s /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@88 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s sha256,sha384,sha512 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:26:48.507 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # connect_authenticate sha256,sha384,sha512 ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 1 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256,sha384,sha512 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.508 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:48.767 nvme0n1 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 0 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 0 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:48.767 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.768 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.027 nvme0n1 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 1 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:49.027 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.028 17:47:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.289 nvme0n1 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 2 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.289 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.549 nvme0n1 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 3 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 3 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.549 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.810 nvme0n1 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 4 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 4 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.810 nvme0n1 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.810 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 0 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:26:50.071 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 0 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.072 nvme0n1 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.072 17:47:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 1 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 1 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.331 nvme0n1 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.331 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 2 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 2 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:50.590 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.591 nvme0n1 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.591 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 3 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:50.850 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 3 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:50.851 nvme0n1 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.851 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 4 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 4 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:51.110 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:51.111 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:51.111 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:51.111 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:51.111 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:51.111 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:26:51.111 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.111 17:47:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.111 nvme0n1 00:26:51.111 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.111 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 0 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 0 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:26:51.370 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.371 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.631 nvme0n1 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 1 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 1 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.631 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.892 nvme0n1 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 2 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 2 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.892 17:47:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.152 nvme0n1 00:26:52.152 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.152 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:52.152 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:52.152 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.152 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.152 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 3 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 3 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.412 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.672 nvme0n1 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.672 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 4 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 4 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.673 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.933 nvme0n1 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 0 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 0 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.933 17:48:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:53.503 nvme0n1 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 1 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:26:53.503 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 1 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:53.504 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.073 nvme0n1 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 2 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 2 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.073 17:48:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.333 nvme0n1 00:26:54.333 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.333 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:54.333 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:54.333 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.333 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.333 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 3 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 3 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.594 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.856 nvme0n1 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.856 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 4 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 4 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:55.117 17:48:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:55.378 nvme0n1 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 0 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 0 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:55.378 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:55.639 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:56.209 nvme0n1 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:56.209 17:48:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 1 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 1 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:56.209 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:56.779 nvme0n1 00:26:56.779 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:56.779 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:56.779 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:56.779 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:56.779 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:56.779 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 2 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 2 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.040 17:48:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:57.612 nvme0n1 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 3 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 3 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.612 17:48:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:58.551 nvme0n1 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:58.551 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 4 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 4 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.552 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.123 nvme0n1 00:26:59.123 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.123 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:59.123 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:59.123 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.123 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.123 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 0 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 0 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.124 17:48:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.384 nvme0n1 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 1 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 1 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.384 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.645 nvme0n1 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 2 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 2 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.645 nvme0n1 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.645 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 3 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 3 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.906 nvme0n1 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:26:59.906 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 4 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 4 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.167 17:48:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.167 nvme0n1 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:00.167 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 0 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 0 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.428 nvme0n1 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.428 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 1 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 1 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.689 nvme0n1 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.689 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 2 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 2 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.949 nvme0n1 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.949 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 3 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:01.209 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 3 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.210 17:48:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.210 nvme0n1 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.210 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 4 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 4 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.470 nvme0n1 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.470 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 0 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 0 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.730 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.991 nvme0n1 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 1 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 1 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.992 17:48:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.253 nvme0n1 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 2 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 2 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.253 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.515 nvme0n1 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 3 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 3 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:02.515 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:27:02.776 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.776 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.777 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.038 nvme0n1 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 4 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 4 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.038 17:48:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.299 nvme0n1 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 0 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 0 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.299 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.873 nvme0n1 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 1 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 1 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.873 17:48:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.135 nvme0n1 00:27:04.135 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.135 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:04.135 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:04.135 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.135 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.135 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.395 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 2 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 2 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.396 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.657 nvme0n1 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 3 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:04.657 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 3 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.658 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.231 nvme0n1 00:27:05.231 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.231 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:05.231 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:05.231 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.231 17:48:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 4 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 4 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.231 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.804 nvme0n1 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 0 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 0 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.804 17:48:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:06.376 nvme0n1 00:27:06.376 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:06.376 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:06.376 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:06.376 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 1 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 1 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:06.377 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.320 nvme0n1 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 2 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 2 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.320 17:48:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.320 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.892 nvme0n1 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 3 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 3 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.892 17:48:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:08.464 nvme0n1 00:27:08.464 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:08.464 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:08.464 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:08.464 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:08.464 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:08.465 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:08.465 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:08.465 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:08.465 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:08.465 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 4 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 4 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:08.726 17:48:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.297 nvme0n1 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 0 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 0 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.297 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.573 nvme0n1 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 1 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 1 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.573 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.884 nvme0n1 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 2 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 2 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.884 nvme0n1 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:09.884 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 3 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 3 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.184 nvme0n1 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.184 17:48:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.184 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.184 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:10.184 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:10.184 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.184 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.184 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.184 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 4 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 4 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.185 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.446 nvme0n1 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 0 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 0 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.446 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.707 nvme0n1 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 1 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 1 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.707 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.968 nvme0n1 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 2 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 2 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.968 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.228 nvme0n1 00:27:11.228 17:48:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 3 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 3 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:11.228 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:11.229 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:11.229 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:11.229 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.229 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.489 nvme0n1 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 4 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 4 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.489 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.749 nvme0n1 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 0 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 0 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:11.749 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.750 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.010 nvme0n1 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:12.010 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 1 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 1 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.272 17:48:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.534 nvme0n1 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 2 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 2 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.534 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.796 nvme0n1 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 3 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 3 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.796 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.057 nvme0n1 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 4 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 4 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.057 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.316 17:48:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.316 nvme0n1 00:27:13.316 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.316 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:13.316 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:13.316 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.316 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.316 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 0 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 0 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.577 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.838 nvme0n1 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 1 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:13.838 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 1 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.098 17:48:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.359 nvme0n1 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 2 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 2 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.359 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.931 nvme0n1 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 3 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 3 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.931 17:48:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:15.502 nvme0n1 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 4 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 4 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:15.502 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:15.763 nvme0n1 00:27:15.763 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 0 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTdmZDQ0YTNkOGM4MmU5YjEwMjM1Yzk5YTk1OGQ0YWH71CQ1: 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:M2Q5ZGE1N2ZiMjczZmZjZGQ1MzBhODdjZTFlZDQzMDFlMGZkOWQyNGNjYTYxMGU4NzY1MjlhZGUxYThiNzE2M5pVmXM=: 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 0 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.024 17:48:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:16.593 nvme0n1 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 1 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 1 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.593 17:48:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:17.536 nvme0n1 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 2 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 2 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:17.536 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:17.537 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:17.537 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:17.537 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:17.537 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:17.537 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.108 nvme0n1 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 3 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YTNlYWFmMzgxMTE3Nzk1ZjU0ZWY4ZmE3ZmQ2ODA4Y2I3MTViNWI4M2FmZjlkODQ0qA3meA==: 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:OTgzMWExY2VmYzFlMTU2ZjZjZTc1ZTBkY2NmMmE3NmaFknG7: 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 3 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.108 17:48:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.677 nvme0n1 00:27:18.677 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.677 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:18.677 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:18.677 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.677 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 4 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:MTM5NTNkZDgzYTE3MGVkMThkMmE2MmRmNGJkM2JlNGU0OTIwMDRhZWNhZDRkNzlkZmU5YWRiOWQ0NThkNDFlNvw+FGY=: 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 4 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.936 17:48:26 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.505 nvme0n1 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@111 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # get_main_ns_ip 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.505 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.765 request: 00:27:19.765 { 00:27:19.765 "name": "nvme0", 00:27:19.765 "trtype": "tcp", 00:27:19.765 "traddr": "10.0.0.1", 00:27:19.765 "adrfam": "ipv4", 00:27:19.765 "trsvcid": "4420", 00:27:19.765 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:27:19.765 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:27:19.765 "prchk_reftag": false, 00:27:19.765 "prchk_guard": false, 00:27:19.765 "hdgst": false, 00:27:19.765 "ddgst": false, 00:27:19.765 "allow_unrecognized_csi": false, 00:27:19.765 "method": "bdev_nvme_attach_controller", 00:27:19.765 "req_id": 1 00:27:19.765 } 00:27:19.765 Got JSON-RPC error response 00:27:19.765 response: 00:27:19.765 { 00:27:19.765 "code": -5, 00:27:19.765 "message": "Input/output error" 00:27:19.765 } 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # rpc_cmd bdev_nvme_get_controllers 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # jq length 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # (( 0 == 0 )) 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # get_main_ns_ip 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.765 request: 00:27:19.765 { 00:27:19.765 "name": "nvme0", 00:27:19.765 "trtype": "tcp", 00:27:19.765 "traddr": "10.0.0.1", 00:27:19.765 "adrfam": "ipv4", 00:27:19.765 "trsvcid": "4420", 00:27:19.765 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:27:19.765 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:27:19.765 "prchk_reftag": false, 00:27:19.765 "prchk_guard": false, 00:27:19.765 "hdgst": false, 00:27:19.765 "ddgst": false, 00:27:19.765 "dhchap_key": "key2", 00:27:19.765 "allow_unrecognized_csi": false, 00:27:19.765 "method": "bdev_nvme_attach_controller", 00:27:19.765 "req_id": 1 00:27:19.765 } 00:27:19.765 Got JSON-RPC error response 00:27:19.765 response: 00:27:19.765 { 00:27:19.765 "code": -5, 00:27:19.765 "message": "Input/output error" 00:27:19.765 } 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:27:19.765 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # rpc_cmd bdev_nvme_get_controllers 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # jq length 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # (( 0 == 0 )) 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # get_main_ns_ip 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.766 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:20.026 request: 00:27:20.026 { 00:27:20.026 "name": "nvme0", 00:27:20.026 "trtype": "tcp", 00:27:20.026 "traddr": "10.0.0.1", 00:27:20.026 "adrfam": "ipv4", 00:27:20.026 "trsvcid": "4420", 00:27:20.026 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:27:20.026 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:27:20.026 "prchk_reftag": false, 00:27:20.026 "prchk_guard": false, 00:27:20.026 "hdgst": false, 00:27:20.026 "ddgst": false, 00:27:20.026 "dhchap_key": "key1", 00:27:20.026 "dhchap_ctrlr_key": "ckey2", 00:27:20.026 "allow_unrecognized_csi": false, 00:27:20.026 "method": "bdev_nvme_attach_controller", 00:27:20.026 "req_id": 1 00:27:20.026 } 00:27:20.026 Got JSON-RPC error response 00:27:20.026 response: 00:27:20.026 { 00:27:20.026 "code": -5, 00:27:20.026 "message": "Input/output error" 00:27:20.026 } 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # get_main_ns_ip 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:20.026 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:20.027 nvme0n1 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@132 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@133 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # rpc_cmd bdev_nvme_get_controllers 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # jq -r '.[].name' 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:20.027 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@136 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:20.287 17:48:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:20.287 request: 00:27:20.287 { 00:27:20.287 "name": "nvme0", 00:27:20.287 "dhchap_key": "key1", 00:27:20.287 "dhchap_ctrlr_key": "ckey2", 00:27:20.287 "method": "bdev_nvme_set_keys", 00:27:20.287 "req_id": 1 00:27:20.287 } 00:27:20.287 Got JSON-RPC error response 00:27:20.287 response: 00:27:20.287 { 00:27:20.287 "code": -13, 00:27:20.287 "message": "Permission denied" 00:27:20.287 } 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 1 != 0 )) 00:27:20.287 17:48:28 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@138 -- # sleep 1s 00:27:21.225 17:48:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:27:21.225 17:48:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:27:21.225 17:48:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:21.225 17:48:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:21.225 17:48:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:21.225 17:48:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 1 != 0 )) 00:27:21.225 17:48:29 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@138 -- # sleep 1s 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 0 != 0 )) 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@141 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:YWNlZjJjNzMzNmVmODc4YmExNjM5NTY2ZDFmYmNmOGE2Y2MzNDJjZmE1MGU0Y2YwNMtwNg==: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:Y2MxOTBlNmExODFlNWIwZGI3Nzc0YWI0NmI5ZmUwMmMxYWRiZDJjMWNjNjVlNmRlyABhDw==: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # get_main_ns_ip 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:22.608 nvme0n1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@146 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NjY1ZjQxMjY3OTZmMzBmZDIzOTk5NDhiNmI0OTVkMzHptdvR: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:MzlkMjE3Njc0MjU2ZDg4ZTJlNmNlMjQyNGY0ZWQ5MTTLU7Q4: 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@147 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:22.608 request: 00:27:22.608 { 00:27:22.608 "name": "nvme0", 00:27:22.608 "dhchap_key": "key2", 00:27:22.608 "dhchap_ctrlr_key": "ckey1", 00:27:22.608 "method": "bdev_nvme_set_keys", 00:27:22.608 "req_id": 1 00:27:22.608 } 00:27:22.608 Got JSON-RPC error response 00:27:22.608 response: 00:27:22.608 { 00:27:22.608 "code": -13, 00:27:22.608 "message": "Permission denied" 00:27:22.608 } 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.608 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:22.609 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.609 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 1 != 0 )) 00:27:22.609 17:48:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@149 -- # sleep 1s 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 0 != 0 )) 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@152 -- # trap - SIGINT SIGTERM EXIT 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@153 -- # cleanup 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@24 -- # nvmftestfini 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@514 -- # nvmfcleanup 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@121 -- # sync 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@124 -- # set +e 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:23.990 rmmod nvme_tcp 00:27:23.990 rmmod nvme_fabrics 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@128 -- # set -e 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@129 -- # return 0 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@515 -- # '[' -n 197856 ']' 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@516 -- # killprocess 197856 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@950 -- # '[' -z 197856 ']' 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@954 -- # kill -0 197856 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@955 -- # uname 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 197856 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@968 -- # echo 'killing process with pid 197856' 00:27:23.990 killing process with pid 197856 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@969 -- # kill 197856 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@974 -- # wait 197856 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@297 -- # iptr 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@789 -- # iptables-save 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@789 -- # iptables-restore 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:23.990 17:48:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:25.901 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:25.901 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@25 -- # rm /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:27:25.901 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@26 -- # rmdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@27 -- # clean_kernel_target 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@710 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 ]] 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@712 -- # echo 0 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@714 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2024-02.io.spdk:cnode0 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@715 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@716 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@719 -- # modules=(/sys/module/nvmet/holders/*) 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@721 -- # modprobe -r nvmet_tcp nvmet 00:27:26.162 17:48:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@724 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:27:29.468 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:27:29.729 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:27:30.302 17:48:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@28 -- # rm -f /tmp/spdk.key-null.9du /tmp/spdk.key-null.uDJ /tmp/spdk.key-sha256.Hy8 /tmp/spdk.key-sha384.OMZ /tmp/spdk.key-sha512.wIx /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log 00:27:30.302 17:48:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:27:33.609 0000:80:01.6 (8086 0b00): Already using the vfio-pci driver 00:27:33.609 0000:80:01.7 (8086 0b00): Already using the vfio-pci driver 00:27:33.609 0000:80:01.4 (8086 0b00): Already using the vfio-pci driver 00:27:33.609 0000:80:01.5 (8086 0b00): Already using the vfio-pci driver 00:27:33.609 0000:80:01.2 (8086 0b00): Already using the vfio-pci driver 00:27:33.609 0000:80:01.3 (8086 0b00): Already using the vfio-pci driver 00:27:33.609 0000:80:01.0 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:80:01.1 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:00:01.6 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:65:00.0 (144d a80a): Already using the vfio-pci driver 00:27:33.610 0000:00:01.7 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:00:01.4 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:00:01.5 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:00:01.2 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:00:01.3 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:00:01.0 (8086 0b00): Already using the vfio-pci driver 00:27:33.610 0000:00:01.1 (8086 0b00): Already using the vfio-pci driver 00:27:34.182 00:27:34.182 real 1m1.150s 00:27:34.182 user 0m54.697s 00:27:34.182 sys 0m16.378s 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:27:34.182 ************************************ 00:27:34.182 END TEST nvmf_auth_host 00:27:34.182 ************************************ 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@32 -- # [[ tcp == \t\c\p ]] 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@33 -- # run_test nvmf_digest /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:34.182 ************************************ 00:27:34.182 START TEST nvmf_digest 00:27:34.182 ************************************ 00:27:34.182 17:48:41 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:27:34.182 * Looking for test storage... 00:27:34.182 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:34.182 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:27:34.182 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1691 -- # lcov --version 00:27:34.182 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # IFS=.-: 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # read -ra ver1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # IFS=.-: 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # read -ra ver2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@338 -- # local 'op=<' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@340 -- # ver1_l=2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@341 -- # ver2_l=1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@344 -- # case "$op" in 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@345 -- # : 1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # decimal 1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # ver1[v]=1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # decimal 2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # ver2[v]=2 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # return 0 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:27:34.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:34.445 --rc genhtml_branch_coverage=1 00:27:34.445 --rc genhtml_function_coverage=1 00:27:34.445 --rc genhtml_legend=1 00:27:34.445 --rc geninfo_all_blocks=1 00:27:34.445 --rc geninfo_unexecuted_blocks=1 00:27:34.445 00:27:34.445 ' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:27:34.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:34.445 --rc genhtml_branch_coverage=1 00:27:34.445 --rc genhtml_function_coverage=1 00:27:34.445 --rc genhtml_legend=1 00:27:34.445 --rc geninfo_all_blocks=1 00:27:34.445 --rc geninfo_unexecuted_blocks=1 00:27:34.445 00:27:34.445 ' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:27:34.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:34.445 --rc genhtml_branch_coverage=1 00:27:34.445 --rc genhtml_function_coverage=1 00:27:34.445 --rc genhtml_legend=1 00:27:34.445 --rc geninfo_all_blocks=1 00:27:34.445 --rc geninfo_unexecuted_blocks=1 00:27:34.445 00:27:34.445 ' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:27:34.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:34.445 --rc genhtml_branch_coverage=1 00:27:34.445 --rc genhtml_function_coverage=1 00:27:34.445 --rc genhtml_legend=1 00:27:34.445 --rc geninfo_all_blocks=1 00:27:34.445 --rc geninfo_unexecuted_blocks=1 00:27:34.445 00:27:34.445 ' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # uname -s 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@15 -- # shopt -s extglob 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@5 -- # export PATH 00:27:34.445 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@51 -- # : 0 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:34.446 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@14 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@15 -- # bperfsock=/var/tmp/bperf.sock 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@16 -- # runtime=2 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@136 -- # [[ tcp != \t\c\p ]] 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@138 -- # nvmftestinit 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@474 -- # prepare_net_devs 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@436 -- # local -g is_hw=no 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@438 -- # remove_spdk_ns 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@309 -- # xtrace_disable 00:27:34.446 17:48:42 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:27:42.587 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # pci_devs=() 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # net_devs=() 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # e810=() 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # local -ga e810 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # x722=() 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # local -ga x722 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # mlx=() 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # local -ga mlx 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:27:42.588 Found 0000:31:00.0 (0x8086 - 0x159b) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:27:42.588 Found 0000:31:00.1 (0x8086 - 0x159b) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:27:42.588 Found net devices under 0000:31:00.0: cvl_0_0 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:27:42.588 Found net devices under 0000:31:00.1: cvl_0_1 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # is_hw=yes 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:42.588 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:42.588 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.597 ms 00:27:42.588 00:27:42.588 --- 10.0.0.2 ping statistics --- 00:27:42.588 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:42.588 rtt min/avg/max/mdev = 0.597/0.597/0.597/0.000 ms 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:42.588 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:42.588 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.199 ms 00:27:42.588 00:27:42.588 --- 10.0.0.1 ping statistics --- 00:27:42.588 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:42.588 rtt min/avg/max/mdev = 0.199/0.199/0.199/0.000 ms 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@448 -- # return 0 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@140 -- # trap cleanup SIGINT SIGTERM EXIT 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@141 -- # [[ 0 -eq 1 ]] 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@145 -- # run_test nvmf_digest_clean run_digest 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:42.588 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:27:42.588 ************************************ 00:27:42.588 START TEST nvmf_digest_clean 00:27:42.589 ************************************ 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1125 -- # run_digest 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@120 -- # local dsa_initiator 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # [[ '' == \d\s\a\_\i\n\i\t\i\a\t\o\r ]] 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # dsa_initiator=false 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@123 -- # tgt_params=("--wait-for-rpc") 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@124 -- # nvmfappstart --wait-for-rpc 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@724 -- # xtrace_disable 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@507 -- # nvmfpid=215466 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@508 -- # waitforlisten 215466 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 215466 ']' 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:42.589 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:42.589 17:48:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:42.589 [2024-10-17 17:48:49.917427] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:27:42.589 [2024-10-17 17:48:49.917496] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:42.589 [2024-10-17 17:48:50.009179] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:42.589 [2024-10-17 17:48:50.063267] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:42.589 [2024-10-17 17:48:50.063323] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:42.589 [2024-10-17 17:48:50.063332] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:42.589 [2024-10-17 17:48:50.063339] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:42.589 [2024-10-17 17:48:50.063345] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:42.589 [2024-10-17 17:48:50.064092] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:42.851 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:42.851 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:27:42.851 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:27:42.851 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@730 -- # xtrace_disable 00:27:42.851 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@125 -- # [[ '' == \d\s\a\_\t\a\r\g\e\t ]] 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@126 -- # common_target_config 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@43 -- # rpc_cmd 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:43.112 null0 00:27:43.112 [2024-10-17 17:48:50.882376] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:43.112 [2024-10-17 17:48:50.906666] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@128 -- # run_bperf randread 4096 128 false 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=215540 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 215540 /var/tmp/bperf.sock 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 215540 ']' 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:27:43.112 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:43.112 17:48:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:43.112 [2024-10-17 17:48:50.967065] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:27:43.113 [2024-10-17 17:48:50.967125] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid215540 ] 00:27:43.374 [2024-10-17 17:48:51.049619] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:43.374 [2024-10-17 17:48:51.102576] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:43.947 17:48:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:43.947 17:48:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:27:43.947 17:48:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:27:43.947 17:48:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:27:43.947 17:48:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:27:44.207 17:48:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:44.207 17:48:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:44.778 nvme0n1 00:27:44.778 17:48:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:27:44.778 17:48:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:27:44.778 Running I/O for 2 seconds... 00:27:46.662 18672.00 IOPS, 72.94 MiB/s [2024-10-17T15:48:54.581Z] 20235.00 IOPS, 79.04 MiB/s 00:27:46.662 Latency(us) 00:27:46.662 [2024-10-17T15:48:54.581Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:46.662 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:27:46.662 nvme0n1 : 2.00 20261.95 79.15 0.00 0.00 6310.07 2771.63 19442.35 00:27:46.662 [2024-10-17T15:48:54.581Z] =================================================================================================================== 00:27:46.662 [2024-10-17T15:48:54.581Z] Total : 20261.95 79.15 0.00 0.00 6310.07 2771.63 19442.35 00:27:46.662 { 00:27:46.662 "results": [ 00:27:46.662 { 00:27:46.662 "job": "nvme0n1", 00:27:46.662 "core_mask": "0x2", 00:27:46.662 "workload": "randread", 00:27:46.662 "status": "finished", 00:27:46.662 "queue_depth": 128, 00:27:46.662 "io_size": 4096, 00:27:46.662 "runtime": 2.003657, 00:27:46.662 "iops": 20261.95102255526, 00:27:46.662 "mibps": 79.14824618185648, 00:27:46.662 "io_failed": 0, 00:27:46.662 "io_timeout": 0, 00:27:46.662 "avg_latency_us": 6310.066006535625, 00:27:46.662 "min_latency_us": 2771.6266666666666, 00:27:46.662 "max_latency_us": 19442.346666666668 00:27:46.662 } 00:27:46.662 ], 00:27:46.662 "core_count": 1 00:27:46.662 } 00:27:46.662 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:27:46.662 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:27:46.662 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:27:46.662 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:27:46.662 | select(.opcode=="crc32c") 00:27:46.662 | "\(.module_name) \(.executed)"' 00:27:46.662 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 215540 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 215540 ']' 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 215540 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 215540 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 215540' 00:27:46.922 killing process with pid 215540 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 215540 00:27:46.922 Received shutdown signal, test time was about 2.000000 seconds 00:27:46.922 00:27:46.922 Latency(us) 00:27:46.922 [2024-10-17T15:48:54.841Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:46.922 [2024-10-17T15:48:54.841Z] =================================================================================================================== 00:27:46.922 [2024-10-17T15:48:54.841Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:46.922 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 215540 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@129 -- # run_bperf randread 131072 16 false 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=216358 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 216358 /var/tmp/bperf.sock 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 216358 ']' 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:27:47.182 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:47.182 17:48:54 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:47.182 [2024-10-17 17:48:54.963430] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:27:47.182 [2024-10-17 17:48:54.963484] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid216358 ] 00:27:47.182 I/O size of 131072 is greater than zero copy threshold (65536). 00:27:47.182 Zero copy mechanism will not be used. 00:27:47.182 [2024-10-17 17:48:55.038131] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:47.182 [2024-10-17 17:48:55.067651] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:48.121 17:48:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:48.121 17:48:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:27:48.121 17:48:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:27:48.121 17:48:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:27:48.121 17:48:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:27:48.121 17:48:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:48.121 17:48:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:48.381 nvme0n1 00:27:48.381 17:48:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:27:48.381 17:48:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:27:48.640 I/O size of 131072 is greater than zero copy threshold (65536). 00:27:48.640 Zero copy mechanism will not be used. 00:27:48.640 Running I/O for 2 seconds... 00:27:50.521 3022.00 IOPS, 377.75 MiB/s [2024-10-17T15:48:58.440Z] 3147.00 IOPS, 393.38 MiB/s 00:27:50.521 Latency(us) 00:27:50.521 [2024-10-17T15:48:58.440Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:50.521 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:27:50.521 nvme0n1 : 2.00 3147.33 393.42 0.00 0.00 5080.40 662.19 7864.32 00:27:50.521 [2024-10-17T15:48:58.440Z] =================================================================================================================== 00:27:50.521 [2024-10-17T15:48:58.440Z] Total : 3147.33 393.42 0.00 0.00 5080.40 662.19 7864.32 00:27:50.521 { 00:27:50.521 "results": [ 00:27:50.521 { 00:27:50.521 "job": "nvme0n1", 00:27:50.521 "core_mask": "0x2", 00:27:50.521 "workload": "randread", 00:27:50.521 "status": "finished", 00:27:50.521 "queue_depth": 16, 00:27:50.521 "io_size": 131072, 00:27:50.521 "runtime": 2.004874, 00:27:50.521 "iops": 3147.329956895047, 00:27:50.521 "mibps": 393.41624461188087, 00:27:50.521 "io_failed": 0, 00:27:50.521 "io_timeout": 0, 00:27:50.521 "avg_latency_us": 5080.397489698891, 00:27:50.521 "min_latency_us": 662.1866666666666, 00:27:50.521 "max_latency_us": 7864.32 00:27:50.521 } 00:27:50.521 ], 00:27:50.521 "core_count": 1 00:27:50.521 } 00:27:50.521 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:27:50.521 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:27:50.521 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:27:50.521 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:27:50.521 | select(.opcode=="crc32c") 00:27:50.521 | "\(.module_name) \(.executed)"' 00:27:50.521 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 216358 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 216358 ']' 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 216358 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 216358 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 216358' 00:27:50.782 killing process with pid 216358 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 216358 00:27:50.782 Received shutdown signal, test time was about 2.000000 seconds 00:27:50.782 00:27:50.782 Latency(us) 00:27:50.782 [2024-10-17T15:48:58.701Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:50.782 [2024-10-17T15:48:58.701Z] =================================================================================================================== 00:27:50.782 [2024-10-17T15:48:58.701Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:50.782 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 216358 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@130 -- # run_bperf randwrite 4096 128 false 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=217186 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 217186 /var/tmp/bperf.sock 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 217186 ']' 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:27:51.043 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:51.043 17:48:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:51.043 [2024-10-17 17:48:58.783950] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:27:51.043 [2024-10-17 17:48:58.784002] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid217186 ] 00:27:51.043 [2024-10-17 17:48:58.859002] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:51.043 [2024-10-17 17:48:58.888402] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:51.986 17:48:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:51.986 17:48:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:27:51.986 17:48:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:27:51.986 17:48:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:27:51.986 17:48:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:27:51.986 17:48:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:51.986 17:48:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:52.247 nvme0n1 00:27:52.247 17:49:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:27:52.247 17:49:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:27:52.247 Running I/O for 2 seconds... 00:27:54.575 30192.00 IOPS, 117.94 MiB/s [2024-10-17T15:49:02.494Z] 30268.00 IOPS, 118.23 MiB/s 00:27:54.575 Latency(us) 00:27:54.575 [2024-10-17T15:49:02.494Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:54.575 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:27:54.575 nvme0n1 : 2.00 30263.22 118.22 0.00 0.00 4222.58 1699.84 14854.83 00:27:54.575 [2024-10-17T15:49:02.494Z] =================================================================================================================== 00:27:54.575 [2024-10-17T15:49:02.494Z] Total : 30263.22 118.22 0.00 0.00 4222.58 1699.84 14854.83 00:27:54.575 { 00:27:54.575 "results": [ 00:27:54.575 { 00:27:54.575 "job": "nvme0n1", 00:27:54.575 "core_mask": "0x2", 00:27:54.575 "workload": "randwrite", 00:27:54.575 "status": "finished", 00:27:54.575 "queue_depth": 128, 00:27:54.575 "io_size": 4096, 00:27:54.575 "runtime": 2.004281, 00:27:54.575 "iops": 30263.22157422038, 00:27:54.575 "mibps": 118.21570927429836, 00:27:54.575 "io_failed": 0, 00:27:54.575 "io_timeout": 0, 00:27:54.575 "avg_latency_us": 4222.583029983293, 00:27:54.575 "min_latency_us": 1699.84, 00:27:54.575 "max_latency_us": 14854.826666666666 00:27:54.575 } 00:27:54.575 ], 00:27:54.575 "core_count": 1 00:27:54.575 } 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:27:54.575 | select(.opcode=="crc32c") 00:27:54.575 | "\(.module_name) \(.executed)"' 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 217186 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 217186 ']' 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 217186 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 217186 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 217186' 00:27:54.575 killing process with pid 217186 00:27:54.575 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 217186 00:27:54.575 Received shutdown signal, test time was about 2.000000 seconds 00:27:54.575 00:27:54.575 Latency(us) 00:27:54.575 [2024-10-17T15:49:02.494Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:54.575 [2024-10-17T15:49:02.495Z] =================================================================================================================== 00:27:54.576 [2024-10-17T15:49:02.495Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:54.576 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 217186 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@131 -- # run_bperf randwrite 131072 16 false 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=217872 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 217872 /var/tmp/bperf.sock 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 217872 ']' 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:27:54.837 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:54.837 17:49:02 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:54.837 [2024-10-17 17:49:02.600516] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:27:54.837 [2024-10-17 17:49:02.600577] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid217872 ] 00:27:54.837 I/O size of 131072 is greater than zero copy threshold (65536). 00:27:54.837 Zero copy mechanism will not be used. 00:27:54.837 [2024-10-17 17:49:02.676227] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:54.837 [2024-10-17 17:49:02.703675] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:55.780 17:49:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:55.780 17:49:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:27:55.780 17:49:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:27:55.780 17:49:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:27:55.780 17:49:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:27:55.780 17:49:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:55.780 17:49:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:27:56.349 nvme0n1 00:27:56.349 17:49:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:27:56.349 17:49:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:27:56.349 I/O size of 131072 is greater than zero copy threshold (65536). 00:27:56.349 Zero copy mechanism will not be used. 00:27:56.349 Running I/O for 2 seconds... 00:27:58.232 3931.00 IOPS, 491.38 MiB/s [2024-10-17T15:49:06.151Z] 4860.50 IOPS, 607.56 MiB/s 00:27:58.232 Latency(us) 00:27:58.232 [2024-10-17T15:49:06.151Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:58.232 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:27:58.232 nvme0n1 : 2.00 4860.46 607.56 0.00 0.00 3287.87 1194.67 14527.15 00:27:58.232 [2024-10-17T15:49:06.151Z] =================================================================================================================== 00:27:58.232 [2024-10-17T15:49:06.151Z] Total : 4860.46 607.56 0.00 0.00 3287.87 1194.67 14527.15 00:27:58.232 { 00:27:58.232 "results": [ 00:27:58.232 { 00:27:58.232 "job": "nvme0n1", 00:27:58.232 "core_mask": "0x2", 00:27:58.232 "workload": "randwrite", 00:27:58.232 "status": "finished", 00:27:58.232 "queue_depth": 16, 00:27:58.232 "io_size": 131072, 00:27:58.232 "runtime": 2.003104, 00:27:58.232 "iops": 4860.456571401185, 00:27:58.232 "mibps": 607.5570714251481, 00:27:58.232 "io_failed": 0, 00:27:58.232 "io_timeout": 0, 00:27:58.232 "avg_latency_us": 3287.865286222953, 00:27:58.232 "min_latency_us": 1194.6666666666667, 00:27:58.232 "max_latency_us": 14527.146666666667 00:27:58.232 } 00:27:58.232 ], 00:27:58.232 "core_count": 1 00:27:58.232 } 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:27:58.492 | select(.opcode=="crc32c") 00:27:58.492 | "\(.module_name) \(.executed)"' 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 217872 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 217872 ']' 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 217872 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:58.492 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 217872 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 217872' 00:27:58.754 killing process with pid 217872 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 217872 00:27:58.754 Received shutdown signal, test time was about 2.000000 seconds 00:27:58.754 00:27:58.754 Latency(us) 00:27:58.754 [2024-10-17T15:49:06.673Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:58.754 [2024-10-17T15:49:06.673Z] =================================================================================================================== 00:27:58.754 [2024-10-17T15:49:06.673Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 217872 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@132 -- # killprocess 215466 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 215466 ']' 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 215466 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 215466 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 215466' 00:27:58.754 killing process with pid 215466 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 215466 00:27:58.754 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 215466 00:27:59.015 00:27:59.015 real 0m16.839s 00:27:59.015 user 0m33.461s 00:27:59.015 sys 0m3.652s 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:27:59.015 ************************************ 00:27:59.015 END TEST nvmf_digest_clean 00:27:59.015 ************************************ 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@147 -- # run_test nvmf_digest_error run_digest_error 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:27:59.015 ************************************ 00:27:59.015 START TEST nvmf_digest_error 00:27:59.015 ************************************ 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1125 -- # run_digest_error 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@102 -- # nvmfappstart --wait-for-rpc 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@724 -- # xtrace_disable 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@507 -- # nvmfpid=218586 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@508 -- # waitforlisten 218586 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 218586 ']' 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:59.015 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:59.015 17:49:06 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:27:59.015 [2024-10-17 17:49:06.827413] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:27:59.015 [2024-10-17 17:49:06.827467] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:59.015 [2024-10-17 17:49:06.915421] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:59.349 [2024-10-17 17:49:06.948625] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:59.349 [2024-10-17 17:49:06.948654] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:59.349 [2024-10-17 17:49:06.948660] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:59.349 [2024-10-17 17:49:06.948665] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:59.349 [2024-10-17 17:49:06.948669] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:59.349 [2024-10-17 17:49:06.949196] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@730 -- # xtrace_disable 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@104 -- # rpc_cmd accel_assign_opc -o crc32c -m error 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:27:59.983 [2024-10-17 17:49:07.667160] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation crc32c will be assigned to module error 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@105 -- # common_target_config 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@43 -- # rpc_cmd 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:27:59.983 null0 00:27:59.983 [2024-10-17 17:49:07.745084] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:59.983 [2024-10-17 17:49:07.769281] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@108 -- # run_bperf_err randread 4096 128 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=218934 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 218934 /var/tmp/bperf.sock 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 218934 ']' 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:27:59.983 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:59.983 17:49:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:27:59.983 [2024-10-17 17:49:07.827719] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:27:59.983 [2024-10-17 17:49:07.827771] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid218934 ] 00:28:00.244 [2024-10-17 17:49:07.903594] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:00.244 [2024-10-17 17:49:07.933331] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:00.815 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:00.815 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:28:00.815 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:00.815 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:01.075 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:28:01.075 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:01.075 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:01.075 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:01.075 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:01.076 17:49:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:01.336 nvme0n1 00:28:01.336 17:49:09 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:28:01.336 17:49:09 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:01.336 17:49:09 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:01.336 17:49:09 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:01.336 17:49:09 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:28:01.336 17:49:09 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:28:01.336 Running I/O for 2 seconds... 00:28:01.597 [2024-10-17 17:49:09.273314] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.273346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:5639 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.273359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.283618] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.283638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:1330 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.283646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.294368] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.294388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:11166 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.294395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.304128] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.304147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:8465 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.304154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.316767] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.316784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:6235 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.316791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.326261] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.326278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:1432 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.326284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.334940] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.334958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:18606 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.334965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.344064] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.344081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:17810 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.344088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.353670] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.353688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:2682 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.353698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:59 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.362150] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.362168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:25494 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.362174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.371127] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.371143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:10884 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.371149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.381015] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.381033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:4854 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.381040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.392315] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.392332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:2278 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.597 [2024-10-17 17:49:09.392338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.597 [2024-10-17 17:49:09.401379] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.597 [2024-10-17 17:49:09.401396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:13230 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.401402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.411862] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.411880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:9339 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.411886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.421066] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.421084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:8535 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.421090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.429906] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.429923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:13650 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.429929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.439821] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.439838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:22801 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.439848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.447635] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.447653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:11806 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.447659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.456999] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.457016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:23162 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.457023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.465939] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.465957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:1187 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.465963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.475526] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.475542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:22426 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.475549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.483486] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.483503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:18729 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.483509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.493258] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.493275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:17367 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.493281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.503227] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.503244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:17643 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.503250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.598 [2024-10-17 17:49:09.512116] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.598 [2024-10-17 17:49:09.512132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:13789 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.598 [2024-10-17 17:49:09.512139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.521613] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.521634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:11144 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.521640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.529569] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.529586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:150 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.529592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.538675] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.538696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20335 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.538703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.548274] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.548291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:791 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.548298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.556225] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.556242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:18774 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.556248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.568116] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.568133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:910 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.568140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.577873] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.577890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:674 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.577896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.585595] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.585612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:13623 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.585618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.595005] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.595022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:4829 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.595029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.603904] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.603921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:9543 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.603928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.612990] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.613007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:4134 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.613014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.622094] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.622112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:6879 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.622118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.630813] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.630830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:25243 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.630836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.639321] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.639339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:8697 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.639345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.648252] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.648269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:9936 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.648275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.656644] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.656661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:11612 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.656667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.665859] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.858 [2024-10-17 17:49:09.665876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:7994 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.858 [2024-10-17 17:49:09.665882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.858 [2024-10-17 17:49:09.675193] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.675210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:11146 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.675221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.683800] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.683817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:16547 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.683824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.692422] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.692439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:25263 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.692446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.701306] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.701323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:1431 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.701329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.709811] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.709827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:3987 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.709833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.718888] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.718905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:4307 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.718911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.728353] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.728370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:12167 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.728376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.736816] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.736833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:19131 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.736839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.746468] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.746485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:1370 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.746491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.755598] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.755617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:20931 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.755623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:94 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.763808] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.763825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:12612 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.763832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:01.859 [2024-10-17 17:49:09.773398] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:01.859 [2024-10-17 17:49:09.773414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:19939 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:01.859 [2024-10-17 17:49:09.773420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.781334] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.781351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:10200 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.781357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.790877] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.790894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:25351 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.790900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:59 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.800276] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.800292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:14325 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.800299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.808083] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.808100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:9852 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.808106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.817521] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.817537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:24499 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.817543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.825546] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.825563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:21346 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.825572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.834610] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.834627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:10328 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.834633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.843776] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.843793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:5663 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.843800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.852876] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.119 [2024-10-17 17:49:09.852892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:19780 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.119 [2024-10-17 17:49:09.852898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.119 [2024-10-17 17:49:09.860846] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.860862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20168 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.860869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.869576] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.869593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:9696 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.869599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.879780] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.879797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:19095 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.879803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.888604] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.888621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:13969 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.888627] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.897372] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.897388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:13403 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.897395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.906497] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.906517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:1194 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.906523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.916133] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.916150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:18403 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.916156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.923904] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.923920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:8531 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.923926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.934115] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.934132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:13424 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.934138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.942969] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.942985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:23429 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.942992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.952811] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.952828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:6699 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.952834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.961461] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.961478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:20721 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.961484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.971140] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.971157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:2700 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.971163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.979968] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.979985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:6642 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.979991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.988315] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.988332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:14285 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.988339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:09.998167] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:09.998184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:20610 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:09.998191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:10.010242] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:10.010260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:22367 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:10.010267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:10.019843] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:10.019873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:12833 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:10.019884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:10.028394] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:10.028413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:851 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:10.028419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.120 [2024-10-17 17:49:10.037645] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.120 [2024-10-17 17:49:10.037663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:5326 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.120 [2024-10-17 17:49:10.037670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.382 [2024-10-17 17:49:10.045819] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.382 [2024-10-17 17:49:10.045836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:25294 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.382 [2024-10-17 17:49:10.045843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.382 [2024-10-17 17:49:10.055011] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.382 [2024-10-17 17:49:10.055028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:1776 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.382 [2024-10-17 17:49:10.055035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.382 [2024-10-17 17:49:10.063694] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.063711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:12664 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.063722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.073335] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.073352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:1611 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.073358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.082167] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.082184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:15853 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.082190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.090162] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.090179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:4591 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.090185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.099378] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.099395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:14122 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.099401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.108570] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.108587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:24356 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.108593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.117144] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.117161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:3328 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.117168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.126214] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.126230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:5470 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.126237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.136571] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.136588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:21378 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.136594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.144560] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.144581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:5420 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.144588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.153540] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.153557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:21880 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.153563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.164260] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.164277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:21035 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.164283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.173351] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.173367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:14797 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.173373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.182652] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.182670] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:19438 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.182676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.191931] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.191948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:15356 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.191955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.200950] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.200967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23031 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.200973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.209901] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.209917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:11141 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.209924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.218861] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.218878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:8865 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.218887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.227829] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.227845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:8796 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.227851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.237040] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.237056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:274 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.237063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.244943] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.244960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:5719 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.244966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 27483.00 IOPS, 107.36 MiB/s [2024-10-17T15:49:10.302Z] [2024-10-17 17:49:10.254838] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.254855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:24106 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.254862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.264580] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.264596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:8606 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.264602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.273906] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.273923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:3193 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.273930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.283058] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.283075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:2590 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.283082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.383 [2024-10-17 17:49:10.290966] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.383 [2024-10-17 17:49:10.290983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:11044 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.383 [2024-10-17 17:49:10.290989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.302176] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.302197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:1013 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.302203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.312953] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.312970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:16350 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.312976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.322706] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.322723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:7909 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.322729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.332327] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.332344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:19412 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.332351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.339879] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.339896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:21188 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.339903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.349551] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.349568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23455 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.349574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.360199] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.360216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:8936 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.360222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.368067] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.368084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:12850 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.368091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.378343] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.378361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:2206 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.378367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.385948] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.385965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:7884 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.385971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.396008] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.396024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:11411 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.396031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.404865] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.404882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:18919 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.404889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.413820] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.413837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:10057 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.413843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.423606] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.423624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:13605 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.423630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.431033] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.431050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:24704 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.431056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.440550] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.645 [2024-10-17 17:49:10.440567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:584 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.645 [2024-10-17 17:49:10.440573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.645 [2024-10-17 17:49:10.449143] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.449159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:6581 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.449166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.458118] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.458135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:21786 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.458144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.467542] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.467560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:8061 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.467566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.477435] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.477452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:24671 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.477459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.487055] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.487071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:6663 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.487077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:74 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.495043] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.495060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:4552 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.495066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.503770] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.503787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:2179 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.503793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.513697] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.513713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:15585 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.513720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.521994] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.522011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:22419 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.522017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.530067] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.530084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:14592 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.530090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.540750] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.540767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:12991 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.540774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.549754] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.549770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:22072 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.549777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.646 [2024-10-17 17:49:10.558961] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.646 [2024-10-17 17:49:10.558978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:23772 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.646 [2024-10-17 17:49:10.558984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.567544] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.567561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:19778 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.567568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.576328] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.576345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:24458 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.576351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.584545] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.584562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:14969 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.584569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.593998] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.594014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:4503 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.594021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.603959] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.603975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:23486 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.603982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.612126] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.612144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:8307 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.612153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.620432] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.620449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:1390 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.620455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:59 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.629750] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.629767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:1189 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.629773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.639510] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.639527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:18002 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.639533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.648940] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.648956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:8500 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.648962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.656663] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.656680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:9868 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.656686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.665344] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.665361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:21019 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.665368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.674499] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.674516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:1421 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.674523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:52 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.684121] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.684138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:5560 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.684144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.693176] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.693195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:25341 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.693201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.701760] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.701777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:20796 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.701784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.711121] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.711138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:24240 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.711144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.719319] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.719336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:9242 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.719342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.728383] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.728401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18326 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.728407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.737212] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.737229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:17019 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.737236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.745784] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.745801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:19883 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.745807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.755581] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.755598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:19112 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.755604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.763918] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.763935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:21218 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.763941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.772921] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.772938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:11646 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.772945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.782278] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.782296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:21276 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.782303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.790974] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.790992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:20955 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.790998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.800401] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.800418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18196 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.909 [2024-10-17 17:49:10.800425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.909 [2024-10-17 17:49:10.809494] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.909 [2024-10-17 17:49:10.809511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:6933 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.910 [2024-10-17 17:49:10.809517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:02.910 [2024-10-17 17:49:10.817807] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:02.910 [2024-10-17 17:49:10.817824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:2642 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:02.910 [2024-10-17 17:49:10.817830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.171 [2024-10-17 17:49:10.827133] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.171 [2024-10-17 17:49:10.827150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:22112 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.171 [2024-10-17 17:49:10.827157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.171 [2024-10-17 17:49:10.836811] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.171 [2024-10-17 17:49:10.836829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:3095 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.171 [2024-10-17 17:49:10.836835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.171 [2024-10-17 17:49:10.844828] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.171 [2024-10-17 17:49:10.844845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:22342 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.171 [2024-10-17 17:49:10.844857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.171 [2024-10-17 17:49:10.853729] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.171 [2024-10-17 17:49:10.853746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:847 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.171 [2024-10-17 17:49:10.853753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.171 [2024-10-17 17:49:10.863854] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.863871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:6930 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.863878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.872015] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.872032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:385 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.872038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.881100] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.881117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:2884 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.881123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.890305] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.890323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:12480 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.890329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.898991] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.899008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:8955 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.899014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.906852] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.906869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:1177 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.906875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.916146] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.916163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:4977 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.916170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.926264] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.926284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:19974 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.926290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.934377] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.934394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:1882 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.934400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.942856] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.942873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:6746 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.942879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.951452] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.951470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:2402 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.951476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.961045] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.961062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:14640 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.961069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.969825] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.969842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23959 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.969848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.978488] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.978505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:16742 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.978511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.987609] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.987626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:2523 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.987632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:10.995444] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:10.995461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:5374 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:10.995471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:59 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.005245] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.005263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:11599 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.005269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.013751] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.013769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:21528 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.013775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.023033] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.023051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:17101 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.023057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.030895] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.030912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:17119 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.030918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.040463] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.040481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:24142 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.040487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.048785] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.048802] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:24183 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.048808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.057824] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.057842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:20831 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.057848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.067842] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.067859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23794 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.067865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:94 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.075402] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.075422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:2703 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.075428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.172 [2024-10-17 17:49:11.084370] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.172 [2024-10-17 17:49:11.084387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:18491 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.172 [2024-10-17 17:49:11.084394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.434 [2024-10-17 17:49:11.093866] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.434 [2024-10-17 17:49:11.093883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:13282 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.434 [2024-10-17 17:49:11.093890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.434 [2024-10-17 17:49:11.102932] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.434 [2024-10-17 17:49:11.102949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:9223 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.434 [2024-10-17 17:49:11.102955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.434 [2024-10-17 17:49:11.111485] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.434 [2024-10-17 17:49:11.111503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:2109 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.434 [2024-10-17 17:49:11.111509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.434 [2024-10-17 17:49:11.121813] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.434 [2024-10-17 17:49:11.121831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:3786 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.434 [2024-10-17 17:49:11.121837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.434 [2024-10-17 17:49:11.129645] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.434 [2024-10-17 17:49:11.129662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:10202 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.129668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.138877] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.138894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:19871 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.138901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.148422] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.148439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:24500 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.148445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.156364] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.156381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:19260 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.156387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.165791] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.165807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:18427 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.165814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:94 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.173973] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.173990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:16870 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.173996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.183670] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.183687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:13696 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.183698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.191816] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.191833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:13402 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.191839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.200173] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.200190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:5672 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.200196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.209822] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.209839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:8639 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.209845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.218438] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.218455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23643 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.218461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.227280] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.227297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:20633 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.227306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.236023] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.236039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:13684 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.236045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.244923] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.244940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:7351 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.244947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 [2024-10-17 17:49:11.254172] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1467cf0) 00:28:03.435 [2024-10-17 17:49:11.254189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:23345 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:03.435 [2024-10-17 17:49:11.254195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:03.435 27936.00 IOPS, 109.12 MiB/s 00:28:03.435 Latency(us) 00:28:03.435 [2024-10-17T15:49:11.354Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:03.435 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:28:03.435 nvme0n1 : 2.00 27953.15 109.19 0.00 0.00 4574.19 2362.03 19551.57 00:28:03.435 [2024-10-17T15:49:11.354Z] =================================================================================================================== 00:28:03.435 [2024-10-17T15:49:11.354Z] Total : 27953.15 109.19 0.00 0.00 4574.19 2362.03 19551.57 00:28:03.435 { 00:28:03.435 "results": [ 00:28:03.435 { 00:28:03.435 "job": "nvme0n1", 00:28:03.435 "core_mask": "0x2", 00:28:03.435 "workload": "randread", 00:28:03.435 "status": "finished", 00:28:03.435 "queue_depth": 128, 00:28:03.435 "io_size": 4096, 00:28:03.435 "runtime": 2.003352, 00:28:03.435 "iops": 27953.150519728933, 00:28:03.435 "mibps": 109.19199421769115, 00:28:03.435 "io_failed": 0, 00:28:03.435 "io_timeout": 0, 00:28:03.435 "avg_latency_us": 4574.1904457142855, 00:28:03.435 "min_latency_us": 2362.0266666666666, 00:28:03.435 "max_latency_us": 19551.573333333334 00:28:03.435 } 00:28:03.435 ], 00:28:03.435 "core_count": 1 00:28:03.435 } 00:28:03.435 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:28:03.435 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:28:03.435 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:28:03.435 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:28:03.435 | .driver_specific 00:28:03.435 | .nvme_error 00:28:03.435 | .status_code 00:28:03.435 | .command_transient_transport_error' 00:28:03.695 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 219 > 0 )) 00:28:03.695 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 218934 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 218934 ']' 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 218934 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 218934 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 218934' 00:28:03.696 killing process with pid 218934 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 218934 00:28:03.696 Received shutdown signal, test time was about 2.000000 seconds 00:28:03.696 00:28:03.696 Latency(us) 00:28:03.696 [2024-10-17T15:49:11.615Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:03.696 [2024-10-17T15:49:11.615Z] =================================================================================================================== 00:28:03.696 [2024-10-17T15:49:11.615Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:03.696 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 218934 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@109 -- # run_bperf_err randread 131072 16 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=219628 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 219628 /var/tmp/bperf.sock 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 219628 ']' 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:28:03.957 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:03.957 17:49:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:03.957 [2024-10-17 17:49:11.674251] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:03.957 [2024-10-17 17:49:11.674303] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid219628 ] 00:28:03.957 I/O size of 131072 is greater than zero copy threshold (65536). 00:28:03.957 Zero copy mechanism will not be used. 00:28:03.957 [2024-10-17 17:49:11.749435] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:03.957 [2024-10-17 17:49:11.778869] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:04.899 17:49:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:05.161 nvme0n1 00:28:05.161 17:49:13 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:28:05.161 17:49:13 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.161 17:49:13 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:05.161 17:49:13 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.161 17:49:13 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:28:05.161 17:49:13 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:28:05.422 I/O size of 131072 is greater than zero copy threshold (65536). 00:28:05.422 Zero copy mechanism will not be used. 00:28:05.422 Running I/O for 2 seconds... 00:28:05.422 [2024-10-17 17:49:13.153479] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.153510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.153519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.164153] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.164174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.164181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.176307] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.176325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.176333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.187394] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.187412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.187419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.199477] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.199496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.199502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.211792] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.211809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.211816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.222031] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.222049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.222055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.232750] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.422 [2024-10-17 17:49:13.232767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.422 [2024-10-17 17:49:13.232774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.422 [2024-10-17 17:49:13.244794] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.244812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.244818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.255962] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.255978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.255985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.268241] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.268258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.268265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.279101] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.279119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.279126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.289748] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.289766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.289777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.301111] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.301129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.301135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.312381] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.312397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.312403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.322218] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.322235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.322241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.423 [2024-10-17 17:49:13.334572] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.423 [2024-10-17 17:49:13.334589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.423 [2024-10-17 17:49:13.334595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.684 [2024-10-17 17:49:13.345157] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.684 [2024-10-17 17:49:13.345174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.684 [2024-10-17 17:49:13.345181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.684 [2024-10-17 17:49:13.354319] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.684 [2024-10-17 17:49:13.354336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.684 [2024-10-17 17:49:13.354343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.684 [2024-10-17 17:49:13.366549] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.684 [2024-10-17 17:49:13.366566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.684 [2024-10-17 17:49:13.366572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.684 [2024-10-17 17:49:13.378872] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.684 [2024-10-17 17:49:13.378889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.684 [2024-10-17 17:49:13.378896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.684 [2024-10-17 17:49:13.391488] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.684 [2024-10-17 17:49:13.391509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.684 [2024-10-17 17:49:13.391515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.684 [2024-10-17 17:49:13.400452] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.684 [2024-10-17 17:49:13.400469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:0 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.400476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.411075] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.411093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.411099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.422171] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.422188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.422195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.432749] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.432766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.432772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.444778] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.444795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.444801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.455140] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.455157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.455163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.465308] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.465327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.465333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.476485] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.476503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.476509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.487033] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.487051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:15488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.487057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.499017] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.499035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:13408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.499042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.508266] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.508285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:10944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.508291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.518710] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.518728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.518735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.530392] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.530410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.530417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.542307] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.542325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.542332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.554242] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.554260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:20512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.554267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.565688] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.565711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:8992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.565718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.577092] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.577110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:10432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.577122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.589180] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.589199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:15968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.589205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.685 [2024-10-17 17:49:13.600982] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.685 [2024-10-17 17:49:13.601001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:2816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.685 [2024-10-17 17:49:13.601007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.611952] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.611970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.611977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.623040] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.623058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.623065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.633461] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.633479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:9536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.633486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.644579] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.644599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:12384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.644605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.656679] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.656702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.656708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.668413] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.668431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:8256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.668437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.678724] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.678745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.678752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.690014] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.690032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.690038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.701644] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.701663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:17312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.701669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.712290] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.712308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.712314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.724270] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.724288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:17408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.724294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.735500] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.735518] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:23680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.735524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.746161] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.746178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.746184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.757774] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.757792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:4768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.757798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.768846] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.768863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.768870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.779032] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.779051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:18208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.779057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.789650] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.789668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:20576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.789675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.801102] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.801121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:12672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.801127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.813633] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.813652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:12576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.813658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.825765] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.825783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.825790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.838501] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.838520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:10304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.838526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.849895] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.849913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:15168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.948 [2024-10-17 17:49:13.849919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:05.948 [2024-10-17 17:49:13.862147] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:05.948 [2024-10-17 17:49:13.862166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:19520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:05.949 [2024-10-17 17:49:13.862172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.875168] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.875187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.875196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.887802] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.887820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:10496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.887826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.899832] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.899850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.899857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.911241] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.911259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:11200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.911266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.922350] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.922368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:4768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.922375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.933214] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.933231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:15136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.933237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.943872] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.943890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:14752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.943896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.954482] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.954499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.954506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.965351] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.965368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.965375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.976089] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.976111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.976117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.986886] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.986904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.986911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:13.997321] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:13.997338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:1952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:13.997344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:14.007421] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:14.007438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:14.007444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:14.018349] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:14.018367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:13120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:14.018373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:14.025974] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:14.025991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:15680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:14.025998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:14.036573] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:14.036591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:15584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:14.036597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:14.048929] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.210 [2024-10-17 17:49:14.048947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.210 [2024-10-17 17:49:14.048953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.210 [2024-10-17 17:49:14.059754] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.211 [2024-10-17 17:49:14.059772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.211 [2024-10-17 17:49:14.059778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.211 [2024-10-17 17:49:14.069744] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.211 [2024-10-17 17:49:14.069762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:12064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.211 [2024-10-17 17:49:14.069769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.211 [2024-10-17 17:49:14.080219] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.211 [2024-10-17 17:49:14.080238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.211 [2024-10-17 17:49:14.080244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.211 [2024-10-17 17:49:14.092177] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.211 [2024-10-17 17:49:14.092195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:3648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.211 [2024-10-17 17:49:14.092201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.211 [2024-10-17 17:49:14.104800] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.211 [2024-10-17 17:49:14.104818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:3008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.211 [2024-10-17 17:49:14.104825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.211 [2024-10-17 17:49:14.116870] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.211 [2024-10-17 17:49:14.116888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.211 [2024-10-17 17:49:14.116894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.129355] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.129374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:4384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.129381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.472 2730.00 IOPS, 341.25 MiB/s [2024-10-17T15:49:14.391Z] [2024-10-17 17:49:14.143370] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.143389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:6528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.143395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.154740] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.154758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:19744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.154764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.162165] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.162187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.162193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.170741] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.170759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:7424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.170765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.178293] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.178311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:13952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.178318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.186810] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.186828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:3200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.186835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.196068] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.196087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:21088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.196093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.203808] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.203825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.203832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.212109] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.212128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:18464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.472 [2024-10-17 17:49:14.212134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.472 [2024-10-17 17:49:14.216578] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.472 [2024-10-17 17:49:14.216596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:15392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.216603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.223416] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.223434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:14976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.223440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.233759] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.233784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.233791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.244518] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.244536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:3968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.244543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.251941] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.251959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:7360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.251965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.257049] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.257067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.257074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.264160] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.264178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.264184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.275746] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.275764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:19328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.275771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.287168] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.287187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:12416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.287193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.299593] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.299612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.299618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.312476] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.312494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:6464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.312504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.323944] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.323963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.323970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.328137] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.328155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:8192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.328161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.333296] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.333314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:21184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.333320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.338360] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.338378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.338384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.346681] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.346703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:1792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.346710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.358260] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.473 [2024-10-17 17:49:14.358278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.473 [2024-10-17 17:49:14.358284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.473 [2024-10-17 17:49:14.370340] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.474 [2024-10-17 17:49:14.370358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.474 [2024-10-17 17:49:14.370364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.474 [2024-10-17 17:49:14.382640] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.474 [2024-10-17 17:49:14.382658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:1984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.474 [2024-10-17 17:49:14.382664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.735 [2024-10-17 17:49:14.395069] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.735 [2024-10-17 17:49:14.395090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.735 [2024-10-17 17:49:14.395097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.406900] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.406918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.406925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.419594] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.419611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.419618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.431620] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.431639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:1216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.431645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.444484] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.444502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:24128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.444508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.458118] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.458136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:21312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.458143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.471059] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.471076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.471083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.480422] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.480441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.480447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.488202] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.488219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:6976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.488226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.496977] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.496995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:11776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.497001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.508326] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.508344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.508350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.520034] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.520052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:5600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.520058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.527982] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.528000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:2144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.528007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.532552] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.532570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:1184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.532576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.540063] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.540081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:18240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.540087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.550377] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.550395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.550402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.561416] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.561434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:5856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.561440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.566396] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.566414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.566423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.570724] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.570743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.570749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.736 [2024-10-17 17:49:14.580186] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.736 [2024-10-17 17:49:14.580204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:10880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.736 [2024-10-17 17:49:14.580210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.584636] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.584655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:12640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.584661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.595333] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.595351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:12640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.595357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.606050] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.606069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.606075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.614440] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.614459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.614465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.621084] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.621102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.621109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.628269] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.628288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.628294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.636583] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.636604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:19552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.636611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.639551] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.639568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:13792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.639574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.737 [2024-10-17 17:49:14.651685] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.737 [2024-10-17 17:49:14.651708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.737 [2024-10-17 17:49:14.651715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.998 [2024-10-17 17:49:14.658319] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.658338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:2272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.658344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.666224] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.666243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.666250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.670614] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.670633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:21184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.670640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.675850] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.675869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.675876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.686246] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.686264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:2432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.686270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.693952] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.693971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:16640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.693978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.699712] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.699731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:7168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.699737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.705113] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.705131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:11584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.705137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.714776] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.714795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:16032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.714801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.726943] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.726963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:8896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.726970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.736661] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.736679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:12064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.736686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.745336] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.745355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:11840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.745361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.752159] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.752177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.752184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.761882] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.761900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:17888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.761907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.770255] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.770273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:20448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.770282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.776941] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.776959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.776965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.783658] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.783676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:11872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.783683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.790738] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.790757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.790763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.798892] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.798910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.798916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.804025] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.804044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:19584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.804050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.808396] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.808414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.808420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.812870] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.812888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:1664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.812894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.821426] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.821443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:4384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.821450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.829459] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.829483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:7584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.829489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.838875] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.838892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:7040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.838899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.846446] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.846465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:16864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.846471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.850763] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.850780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:06.999 [2024-10-17 17:49:14.850787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:06.999 [2024-10-17 17:49:14.858089] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:06.999 [2024-10-17 17:49:14.858107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:11072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.000 [2024-10-17 17:49:14.858113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.000 [2024-10-17 17:49:14.867685] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.000 [2024-10-17 17:49:14.867709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:13760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.000 [2024-10-17 17:49:14.867715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.000 [2024-10-17 17:49:14.875026] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.000 [2024-10-17 17:49:14.875044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:5760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.000 [2024-10-17 17:49:14.875050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.000 [2024-10-17 17:49:14.883533] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.000 [2024-10-17 17:49:14.883551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:9088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.000 [2024-10-17 17:49:14.883558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.000 [2024-10-17 17:49:14.892091] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.000 [2024-10-17 17:49:14.892109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.000 [2024-10-17 17:49:14.892115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.000 [2024-10-17 17:49:14.899042] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.000 [2024-10-17 17:49:14.899060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:5472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.000 [2024-10-17 17:49:14.899066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.000 [2024-10-17 17:49:14.909319] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.000 [2024-10-17 17:49:14.909337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.000 [2024-10-17 17:49:14.909343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.260 [2024-10-17 17:49:14.919752] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.260 [2024-10-17 17:49:14.919771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.260 [2024-10-17 17:49:14.919777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.260 [2024-10-17 17:49:14.930471] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.260 [2024-10-17 17:49:14.930488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:12736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.260 [2024-10-17 17:49:14.930495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.260 [2024-10-17 17:49:14.935721] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.935739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.935745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.943172] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.943190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:19776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.943197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.947565] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.947582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:13920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.947588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.955375] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.955393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:12960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.955399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.960735] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.960753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:15264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.960762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.969756] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.969774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:10560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.969781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.979629] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.979648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:2368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.979655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.983945] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.983962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:12224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.983969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.989556] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.989575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.989581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:14.997624] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:14.997642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:21056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:14.997649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.002671] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.002689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:13600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.002701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.010363] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.010380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:20480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.010387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.017432] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.017450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:6560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.017456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.022516] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.022534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:13856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.022541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.027254] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.027272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.027279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.034782] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.034799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:21664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.034806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.040490] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.040509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:10624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.040515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.044990] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.045008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.045015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.050234] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.050252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.050258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.059818] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.059836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:20544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.059842] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.070665] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.070683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:24320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.070689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.076849] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.076867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:8352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.076877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.087572] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.087589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:7840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.087595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.092592] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.092611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.092617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.099857] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.099875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.099882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.111366] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.261 [2024-10-17 17:49:15.111384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:9472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.261 [2024-10-17 17:49:15.111391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:07.261 [2024-10-17 17:49:15.123061] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.262 [2024-10-17 17:49:15.123079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:23712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.262 [2024-10-17 17:49:15.123085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:07.262 [2024-10-17 17:49:15.134282] nvme_tcp.c:1470:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xca83f0) 00:28:07.262 [2024-10-17 17:49:15.134300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:6784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:07.262 [2024-10-17 17:49:15.134306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:07.262 3247.00 IOPS, 405.88 MiB/s 00:28:07.262 Latency(us) 00:28:07.262 [2024-10-17T15:49:15.181Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:07.262 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:28:07.262 nvme0n1 : 2.00 3250.58 406.32 0.00 0.00 4919.56 771.41 15182.51 00:28:07.262 [2024-10-17T15:49:15.181Z] =================================================================================================================== 00:28:07.262 [2024-10-17T15:49:15.181Z] Total : 3250.58 406.32 0.00 0.00 4919.56 771.41 15182.51 00:28:07.262 { 00:28:07.262 "results": [ 00:28:07.262 { 00:28:07.262 "job": "nvme0n1", 00:28:07.262 "core_mask": "0x2", 00:28:07.262 "workload": "randread", 00:28:07.262 "status": "finished", 00:28:07.262 "queue_depth": 16, 00:28:07.262 "io_size": 131072, 00:28:07.262 "runtime": 2.002719, 00:28:07.262 "iops": 3250.5808353543357, 00:28:07.262 "mibps": 406.32260441929196, 00:28:07.262 "io_failed": 0, 00:28:07.262 "io_timeout": 0, 00:28:07.262 "avg_latency_us": 4919.556849974399, 00:28:07.262 "min_latency_us": 771.4133333333333, 00:28:07.262 "max_latency_us": 15182.506666666666 00:28:07.262 } 00:28:07.262 ], 00:28:07.262 "core_count": 1 00:28:07.262 } 00:28:07.262 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:28:07.262 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:28:07.262 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:28:07.262 | .driver_specific 00:28:07.262 | .nvme_error 00:28:07.262 | .status_code 00:28:07.262 | .command_transient_transport_error' 00:28:07.262 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 209 > 0 )) 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 219628 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 219628 ']' 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 219628 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 219628 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 219628' 00:28:07.522 killing process with pid 219628 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 219628 00:28:07.522 Received shutdown signal, test time was about 2.000000 seconds 00:28:07.522 00:28:07.522 Latency(us) 00:28:07.522 [2024-10-17T15:49:15.441Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:07.522 [2024-10-17T15:49:15.441Z] =================================================================================================================== 00:28:07.522 [2024-10-17T15:49:15.441Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:07.522 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 219628 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@114 -- # run_bperf_err randwrite 4096 128 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=220305 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 220305 /var/tmp/bperf.sock 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 220305 ']' 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:28:07.783 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:07.783 17:49:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:07.783 [2024-10-17 17:49:15.575487] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:07.783 [2024-10-17 17:49:15.575545] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid220305 ] 00:28:07.783 [2024-10-17 17:49:15.652236] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:07.783 [2024-10-17 17:49:15.681613] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:08.725 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:08.726 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:08.986 nvme0n1 00:28:08.986 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:28:08.986 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:08.986 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:08.986 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:08.986 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:28:08.986 17:49:16 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:28:09.247 Running I/O for 2 seconds... 00:28:09.247 [2024-10-17 17:49:16.954497] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f3a28 00:28:09.247 [2024-10-17 17:49:16.955337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:24813 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:16.955365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:16.963174] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f31b8 00:28:09.247 [2024-10-17 17:49:16.963988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:18496 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:16.964010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:16.971664] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f3a28 00:28:09.247 [2024-10-17 17:49:16.972480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:395 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:16.972496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:16.980157] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f31b8 00:28:09.247 [2024-10-17 17:49:16.980983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:14201 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:16.981000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:16.988651] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f3a28 00:28:09.247 [2024-10-17 17:49:16.989477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:17083 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:16.989494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:16.997438] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e3060 00:28:09.247 [2024-10-17 17:49:16.998139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:9876 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:16.998155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:17.006227] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.247 [2024-10-17 17:49:17.007130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:12336 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:17.007147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:17.014753] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.247 [2024-10-17 17:49:17.015648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:17966 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:17.015664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:17.023248] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.247 [2024-10-17 17:49:17.024138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:20653 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:17.024154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.247 [2024-10-17 17:49:17.031734] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.247 [2024-10-17 17:49:17.032647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:4432 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.247 [2024-10-17 17:49:17.032662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.040237] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.248 [2024-10-17 17:49:17.041142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:20861 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.041158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.048722] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.248 [2024-10-17 17:49:17.049616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:23645 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.049631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.057209] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.248 [2024-10-17 17:49:17.058110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:13529 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.058125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.065664] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.248 [2024-10-17 17:49:17.066576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:772 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.066592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.074125] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.248 [2024-10-17 17:49:17.075025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:22365 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.075041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.082579] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.248 [2024-10-17 17:49:17.083482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:11567 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.083498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.091068] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.248 [2024-10-17 17:49:17.091957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:1150 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.091973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.099533] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.248 [2024-10-17 17:49:17.100407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:8944 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.100423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.108042] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.248 [2024-10-17 17:49:17.108956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:24045 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.108972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.116509] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.248 [2024-10-17 17:49:17.117415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:22393 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.117431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.124953] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.248 [2024-10-17 17:49:17.125849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:25106 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.125865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.133408] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.248 [2024-10-17 17:49:17.134265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:18179 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.134281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.141892] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.248 [2024-10-17 17:49:17.142791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:12663 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.142807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.150357] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.248 [2024-10-17 17:49:17.151257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:1987 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.151274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:74 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.248 [2024-10-17 17:49:17.158826] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.248 [2024-10-17 17:49:17.159719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:10041 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.248 [2024-10-17 17:49:17.159735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.520 [2024-10-17 17:49:17.167268] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.521 [2024-10-17 17:49:17.168170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:12764 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.521 [2024-10-17 17:49:17.168186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.521 [2024-10-17 17:49:17.175735] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.521 [2024-10-17 17:49:17.176634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:20599 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.521 [2024-10-17 17:49:17.176649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.521 [2024-10-17 17:49:17.184205] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.521 [2024-10-17 17:49:17.185100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:15944 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.521 [2024-10-17 17:49:17.185119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.521 [2024-10-17 17:49:17.192696] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.521 [2024-10-17 17:49:17.193588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:8901 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.521 [2024-10-17 17:49:17.193604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.521 [2024-10-17 17:49:17.201166] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.521 [2024-10-17 17:49:17.202039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:13516 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.521 [2024-10-17 17:49:17.202055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.521 [2024-10-17 17:49:17.209632] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.521 [2024-10-17 17:49:17.210511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:22936 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.521 [2024-10-17 17:49:17.210527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.521 [2024-10-17 17:49:17.218085] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.521 [2024-10-17 17:49:17.218964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:8426 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.521 [2024-10-17 17:49:17.218979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.521 [2024-10-17 17:49:17.226548] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.522 [2024-10-17 17:49:17.227440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23226 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.522 [2024-10-17 17:49:17.227456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.522 [2024-10-17 17:49:17.235013] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.522 [2024-10-17 17:49:17.235914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:1489 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.522 [2024-10-17 17:49:17.235930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.522 [2024-10-17 17:49:17.243503] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.522 [2024-10-17 17:49:17.244361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:11703 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.522 [2024-10-17 17:49:17.244377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.522 [2024-10-17 17:49:17.251961] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.522 [2024-10-17 17:49:17.252848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:1349 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.522 [2024-10-17 17:49:17.252864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.522 [2024-10-17 17:49:17.260450] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.522 [2024-10-17 17:49:17.261349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:15948 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.522 [2024-10-17 17:49:17.261364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.522 [2024-10-17 17:49:17.268909] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.522 [2024-10-17 17:49:17.269804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:5530 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.523 [2024-10-17 17:49:17.269820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.523 [2024-10-17 17:49:17.277391] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.523 [2024-10-17 17:49:17.278306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:17884 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.523 [2024-10-17 17:49:17.278322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.523 [2024-10-17 17:49:17.285865] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.523 [2024-10-17 17:49:17.286770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:20127 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.523 [2024-10-17 17:49:17.286786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.523 [2024-10-17 17:49:17.294325] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.523 [2024-10-17 17:49:17.295224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:3038 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.523 [2024-10-17 17:49:17.295240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.523 [2024-10-17 17:49:17.302776] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.523 [2024-10-17 17:49:17.303670] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:6537 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.523 [2024-10-17 17:49:17.303685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.311241] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.525 [2024-10-17 17:49:17.312134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:1351 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.525 [2024-10-17 17:49:17.312150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.319717] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.525 [2024-10-17 17:49:17.320601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:7150 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.525 [2024-10-17 17:49:17.320617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.328185] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.525 [2024-10-17 17:49:17.329078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:12036 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.525 [2024-10-17 17:49:17.329094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:74 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.336627] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.525 [2024-10-17 17:49:17.337530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:428 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.525 [2024-10-17 17:49:17.337546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.345108] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.525 [2024-10-17 17:49:17.345986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:8033 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.525 [2024-10-17 17:49:17.346001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.353568] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.525 [2024-10-17 17:49:17.354458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:25052 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.525 [2024-10-17 17:49:17.354473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.362037] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.525 [2024-10-17 17:49:17.362944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:22144 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.525 [2024-10-17 17:49:17.362960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.525 [2024-10-17 17:49:17.370494] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.525 [2024-10-17 17:49:17.371394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:21186 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.371410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.526 [2024-10-17 17:49:17.378973] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.526 [2024-10-17 17:49:17.379884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:7622 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.379900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.526 [2024-10-17 17:49:17.387429] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.526 [2024-10-17 17:49:17.388321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:11454 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.388337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.526 [2024-10-17 17:49:17.395893] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.526 [2024-10-17 17:49:17.396757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:5704 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.396774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.526 [2024-10-17 17:49:17.404324] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.526 [2024-10-17 17:49:17.405214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:15419 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.405230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.526 [2024-10-17 17:49:17.412812] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.526 [2024-10-17 17:49:17.413709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:25154 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.413725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.526 [2024-10-17 17:49:17.421279] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:09.526 [2024-10-17 17:49:17.422168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:21149 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.422184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.526 [2024-10-17 17:49:17.429779] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:09.526 [2024-10-17 17:49:17.430674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:1406 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.526 [2024-10-17 17:49:17.430690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.438258] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5378 00:28:09.787 [2024-10-17 17:49:17.439160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:4662 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.439176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.446727] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaab8 00:28:09.787 [2024-10-17 17:49:17.447621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:1231 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.447636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.455189] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed4e8 00:28:09.787 [2024-10-17 17:49:17.456100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:306 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.456116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.464177] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fc128 00:28:09.787 [2024-10-17 17:49:17.465035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:2166 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.465052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.473923] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f0350 00:28:09.787 [2024-10-17 17:49:17.475476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:1924 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.475491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.479997] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fb048 00:28:09.787 [2024-10-17 17:49:17.480737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13721 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.480757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.488584] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166df988 00:28:09.787 [2024-10-17 17:49:17.489295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:15329 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.489311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.497033] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e0a68 00:28:09.787 [2024-10-17 17:49:17.497764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:4786 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.497779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.505480] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fd640 00:28:09.787 [2024-10-17 17:49:17.506227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:6052 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.506244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.513937] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ec408 00:28:09.787 [2024-10-17 17:49:17.514685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:19197 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.787 [2024-10-17 17:49:17.514704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.787 [2024-10-17 17:49:17.522416] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f9f68 00:28:09.787 [2024-10-17 17:49:17.523159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:7620 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.523175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.530849] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f8e88 00:28:09.788 [2024-10-17 17:49:17.531595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:17639 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.531611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.539283] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f7da8 00:28:09.788 [2024-10-17 17:49:17.540057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:16580 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.540073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.547730] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166de470 00:28:09.788 [2024-10-17 17:49:17.548498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:20357 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.548513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.556182] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f46d0 00:28:09.788 [2024-10-17 17:49:17.556947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:6050 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.556962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.564631] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f57b0 00:28:09.788 [2024-10-17 17:49:17.565398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:20217 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.565414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.573075] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f6890 00:28:09.788 [2024-10-17 17:49:17.573837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:14789 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.573852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.581527] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f35f0 00:28:09.788 [2024-10-17 17:49:17.582300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:8172 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.582316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.589962] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f2510 00:28:09.788 [2024-10-17 17:49:17.590709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:9714 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.590725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.598411] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f1430 00:28:09.788 [2024-10-17 17:49:17.599181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:506 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.599197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.606898] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e8d30 00:28:09.788 [2024-10-17 17:49:17.607619] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:4151 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.607634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.615369] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fa3a0 00:28:09.788 [2024-10-17 17:49:17.616117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:21792 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.616133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.623852] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fb480 00:28:09.788 [2024-10-17 17:49:17.624602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:14044 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.624618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.632295] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dfdc0 00:28:09.788 [2024-10-17 17:49:17.633028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:7726 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.633044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.640779] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e0ea0 00:28:09.788 [2024-10-17 17:49:17.641541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:16983 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.641556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.649251] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166feb58 00:28:09.788 [2024-10-17 17:49:17.650022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:13882 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.650038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.657699] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e73e0 00:28:09.788 [2024-10-17 17:49:17.658451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:11319 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.658467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.666168] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f92c0 00:28:09.788 [2024-10-17 17:49:17.666916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:16649 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.666932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.674990] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ef270 00:28:09.788 [2024-10-17 17:49:17.675809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:11542 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.675825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.683464] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:09.788 [2024-10-17 17:49:17.684326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:20102 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.684342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.691934] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ef270 00:28:09.788 [2024-10-17 17:49:17.692752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:17802 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.692769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:09.788 [2024-10-17 17:49:17.700528] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ee190 00:28:09.788 [2024-10-17 17:49:17.701395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:2863 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:09.788 [2024-10-17 17:49:17.701413] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.049 [2024-10-17 17:49:17.708992] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed0b0 00:28:10.049 [2024-10-17 17:49:17.709855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:9748 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.049 [2024-10-17 17:49:17.709871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.049 [2024-10-17 17:49:17.717439] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f4f40 00:28:10.050 [2024-10-17 17:49:17.718263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:24449 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.718279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.725892] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f6020 00:28:10.050 [2024-10-17 17:49:17.726757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:4623 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.726773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.734340] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fe2e8 00:28:10.050 [2024-10-17 17:49:17.735225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:13938 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.735241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.742823] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fc998 00:28:10.050 [2024-10-17 17:49:17.743682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:15776 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.743701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.751285] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fb8b8 00:28:10.050 [2024-10-17 17:49:17.752158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:22229 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.752174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.759757] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e4de8 00:28:10.050 [2024-10-17 17:49:17.760634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:4779 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.760650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.768213] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e1f80 00:28:10.050 [2024-10-17 17:49:17.769087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:11221 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.769103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.776653] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e3060 00:28:10.050 [2024-10-17 17:49:17.777555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:186 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.777571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.785126] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e4140 00:28:10.050 [2024-10-17 17:49:17.785959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:22622 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.785975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.793572] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f0350 00:28:10.050 [2024-10-17 17:49:17.794447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:22518 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.794463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.802028] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f1ca0 00:28:10.050 [2024-10-17 17:49:17.802874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:397 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.802891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.810456] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f3e60 00:28:10.050 [2024-10-17 17:49:17.811332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:7953 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.811348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.818911] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eea00 00:28:10.050 [2024-10-17 17:49:17.819766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:5159 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.819783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.827344] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eff18 00:28:10.050 [2024-10-17 17:49:17.828207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:17464 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.828223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.835882] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eee38 00:28:10.050 [2024-10-17 17:49:17.836772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:12542 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.836787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.844371] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166edd58 00:28:10.050 [2024-10-17 17:49:17.845252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:260 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.845268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.852816] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:10.050 [2024-10-17 17:49:17.853702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:21612 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.853717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.861248] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f5be8 00:28:10.050 [2024-10-17 17:49:17.862088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:19817 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.862104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.869670] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f2d80 00:28:10.050 [2024-10-17 17:49:17.870546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:1211 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.870562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.878137] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fe720 00:28:10.050 [2024-10-17 17:49:17.879000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:14756 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.879017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.886588] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fc560 00:28:10.050 [2024-10-17 17:49:17.887458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:1945 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.887474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.895058] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e5ec8 00:28:10.050 [2024-10-17 17:49:17.895940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:17199 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.895956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.903505] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e1b48 00:28:10.050 [2024-10-17 17:49:17.904373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:13959 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.904390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.911961] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e2c28 00:28:10.050 [2024-10-17 17:49:17.912822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:3247 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.912838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.920419] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb760 00:28:10.050 [2024-10-17 17:49:17.921283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:3657 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.921301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.928874] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e99d8 00:28:10.050 [2024-10-17 17:49:17.929756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:22754 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.929772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.937348] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f0ff8 00:28:10.050 [2024-10-17 17:49:17.938226] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:6485 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.938243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:28:10.050 29833.00 IOPS, 116.54 MiB/s [2024-10-17T15:49:17.969Z] [2024-10-17 17:49:17.946027] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dece0 00:28:10.050 [2024-10-17 17:49:17.947336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:1173 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.050 [2024-10-17 17:49:17.947353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:28:10.050 [2024-10-17 17:49:17.953704] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166df988 00:28:10.050 [2024-10-17 17:49:17.954447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:7495 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.051 [2024-10-17 17:49:17.954463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:28:10.051 [2024-10-17 17:49:17.962170] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dece0 00:28:10.051 [2024-10-17 17:49:17.962894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:25483 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.051 [2024-10-17 17:49:17.962910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:17.970779] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eee38 00:28:10.312 [2024-10-17 17:49:17.971539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:4938 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:17.971555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:17.979256] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166edd58 00:28:10.312 [2024-10-17 17:49:17.980023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:11381 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:17.980039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:17.987740] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:10.312 [2024-10-17 17:49:17.988484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:16271 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:17.988501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:17.996196] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f57b0 00:28:10.312 [2024-10-17 17:49:17.996926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:21472 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:17.996942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.004646] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e5220 00:28:10.312 [2024-10-17 17:49:18.005356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:24504 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.005372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.013097] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e7c50 00:28:10.312 [2024-10-17 17:49:18.013845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12831 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.013862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.021543] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e6b70 00:28:10.312 [2024-10-17 17:49:18.022312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:8268 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.022328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.029995] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaef0 00:28:10.312 [2024-10-17 17:49:18.030741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:15432 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.030756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.038434] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f96f8 00:28:10.312 [2024-10-17 17:49:18.039202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:6435 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.039218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.046882] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f8618 00:28:10.312 [2024-10-17 17:49:18.047648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:22547 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.047663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.055339] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ddc00 00:28:10.312 [2024-10-17 17:49:18.056113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:8410 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.056129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.063812] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166df988 00:28:10.312 [2024-10-17 17:49:18.064566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:1755 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.064582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.312 [2024-10-17 17:49:18.072250] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e0a68 00:28:10.312 [2024-10-17 17:49:18.072985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:10047 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.312 [2024-10-17 17:49:18.073001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.080713] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fd640 00:28:10.313 [2024-10-17 17:49:18.081456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:1921 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.081472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.089147] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ec408 00:28:10.313 [2024-10-17 17:49:18.089899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:5829 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.089915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.097590] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f4b08 00:28:10.313 [2024-10-17 17:49:18.098343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:16226 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.098359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.106064] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166efae0 00:28:10.313 [2024-10-17 17:49:18.106814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:23792 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.106830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.114512] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e3d08 00:28:10.313 [2024-10-17 17:49:18.115261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:14919 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.115277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.123003] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed920 00:28:10.313 [2024-10-17 17:49:18.123752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:10695 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.123767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.131444] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ec840 00:28:10.313 [2024-10-17 17:49:18.132204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:19795 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.132221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.139905] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f6020 00:28:10.313 [2024-10-17 17:49:18.140628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:22087 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.140646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.148347] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e8088 00:28:10.313 [2024-10-17 17:49:18.149097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:21748 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.149113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.156804] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e6fa8 00:28:10.313 [2024-10-17 17:49:18.157555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:10702 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.157571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.165255] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:10.313 [2024-10-17 17:49:18.165966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16853 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.165981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.173689] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f9b30 00:28:10.313 [2024-10-17 17:49:18.174441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:11915 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.174457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.182143] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f8a50 00:28:10.313 [2024-10-17 17:49:18.182857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:15562 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.182873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.190575] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f7970 00:28:10.313 [2024-10-17 17:49:18.191329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:22771 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.191345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.199023] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166de8a8 00:28:10.313 [2024-10-17 17:49:18.199772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:10689 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.199788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.207471] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dfdc0 00:28:10.313 [2024-10-17 17:49:18.208221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:15917 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.208237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.215921] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e0ea0 00:28:10.313 [2024-10-17 17:49:18.216685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:3509 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.216704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.313 [2024-10-17 17:49:18.224347] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166feb58 00:28:10.313 [2024-10-17 17:49:18.225096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:19842 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.313 [2024-10-17 17:49:18.225112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.232793] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e73e0 00:28:10.575 [2024-10-17 17:49:18.233538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:24478 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.233554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.241257] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dece0 00:28:10.575 [2024-10-17 17:49:18.241970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:1233 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.241986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.249731] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eee38 00:28:10.575 [2024-10-17 17:49:18.250477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:16473 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.250492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.258200] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166edd58 00:28:10.575 [2024-10-17 17:49:18.258971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:20471 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.258987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.266642] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:10.575 [2024-10-17 17:49:18.267406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:2467 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.267423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.275079] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f57b0 00:28:10.575 [2024-10-17 17:49:18.275828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:13130 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.275844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.283512] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e5220 00:28:10.575 [2024-10-17 17:49:18.284266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:20482 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.284281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.291950] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e7c50 00:28:10.575 [2024-10-17 17:49:18.292715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:9470 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.292731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.300389] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e6b70 00:28:10.575 [2024-10-17 17:49:18.301102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:13240 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.301117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.308826] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaef0 00:28:10.575 [2024-10-17 17:49:18.309572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6557 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.309588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.317241] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f96f8 00:28:10.575 [2024-10-17 17:49:18.317963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:9392 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.317979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.325656] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f8618 00:28:10.575 [2024-10-17 17:49:18.326407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:9074 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.326423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.334095] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ddc00 00:28:10.575 [2024-10-17 17:49:18.334837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:21875 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.575 [2024-10-17 17:49:18.334854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.575 [2024-10-17 17:49:18.342552] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166df988 00:28:10.575 [2024-10-17 17:49:18.343321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:19708 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.343337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.351001] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e0a68 00:28:10.576 [2024-10-17 17:49:18.351765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:17684 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.351780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.359438] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fd640 00:28:10.576 [2024-10-17 17:49:18.360194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23134 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.360213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.367858] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ec408 00:28:10.576 [2024-10-17 17:49:18.368607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:22559 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.368623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.376280] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f4b08 00:28:10.576 [2024-10-17 17:49:18.377029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:21964 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.377045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.384733] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166efae0 00:28:10.576 [2024-10-17 17:49:18.385479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:21644 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.385494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.393238] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e3d08 00:28:10.576 [2024-10-17 17:49:18.394009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:766 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.394025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.401699] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed920 00:28:10.576 [2024-10-17 17:49:18.402423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:10415 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.402439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.410129] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ec840 00:28:10.576 [2024-10-17 17:49:18.410900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:22943 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.410915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.418558] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f6020 00:28:10.576 [2024-10-17 17:49:18.419299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:9511 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.419315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.427007] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e8088 00:28:10.576 [2024-10-17 17:49:18.427767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:9917 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.427782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.435468] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e6fa8 00:28:10.576 [2024-10-17 17:49:18.436243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:24015 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.436259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.443930] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:10.576 [2024-10-17 17:49:18.444675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:11577 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.444693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.452357] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f9b30 00:28:10.576 [2024-10-17 17:49:18.453121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:6568 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.453137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.460788] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f8a50 00:28:10.576 [2024-10-17 17:49:18.461533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:14647 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.461548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.469401] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f7970 00:28:10.576 [2024-10-17 17:49:18.470144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:8659 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.470160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.477858] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166de8a8 00:28:10.576 [2024-10-17 17:49:18.478594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:2702 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.478609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.576 [2024-10-17 17:49:18.486302] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dfdc0 00:28:10.576 [2024-10-17 17:49:18.487072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:13178 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.576 [2024-10-17 17:49:18.487088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.837 [2024-10-17 17:49:18.494755] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e0ea0 00:28:10.837 [2024-10-17 17:49:18.495519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:25145 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.837 [2024-10-17 17:49:18.495536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.837 [2024-10-17 17:49:18.503196] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166feb58 00:28:10.837 [2024-10-17 17:49:18.503958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:20676 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.837 [2024-10-17 17:49:18.503973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.837 [2024-10-17 17:49:18.511636] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e73e0 00:28:10.837 [2024-10-17 17:49:18.512411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:4801 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.837 [2024-10-17 17:49:18.512427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.837 [2024-10-17 17:49:18.520087] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dece0 00:28:10.837 [2024-10-17 17:49:18.520856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:6335 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.837 [2024-10-17 17:49:18.520872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.837 [2024-10-17 17:49:18.528537] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eee38 00:28:10.837 [2024-10-17 17:49:18.529285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:1979 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.837 [2024-10-17 17:49:18.529301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.837 [2024-10-17 17:49:18.537000] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166edd58 00:28:10.838 [2024-10-17 17:49:18.537763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:14937 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.537778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.545418] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ecc78 00:28:10.838 [2024-10-17 17:49:18.546183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:14375 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.546200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.553863] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f57b0 00:28:10.838 [2024-10-17 17:49:18.554633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:22086 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.554649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.562287] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e5220 00:28:10.838 [2024-10-17 17:49:18.563030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:5099 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.563046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.570745] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e7c50 00:28:10.838 [2024-10-17 17:49:18.571492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:25594 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.571507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.579196] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e6b70 00:28:10.838 [2024-10-17 17:49:18.579968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:15425 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.579987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.587645] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eaef0 00:28:10.838 [2024-10-17 17:49:18.588414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:8591 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.588431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.596077] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f96f8 00:28:10.838 [2024-10-17 17:49:18.596819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:22308 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.596836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.604509] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f8618 00:28:10.838 [2024-10-17 17:49:18.605263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:5823 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.605279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.612949] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ddc00 00:28:10.838 [2024-10-17 17:49:18.613698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:15811 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.613715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.621401] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166df988 00:28:10.838 [2024-10-17 17:49:18.622154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:9929 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.622170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.629860] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e0a68 00:28:10.838 [2024-10-17 17:49:18.630609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:13809 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.630624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.638285] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fd640 00:28:10.838 [2024-10-17 17:49:18.639029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:8558 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.639045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.646723] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ec408 00:28:10.838 [2024-10-17 17:49:18.647487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:2707 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.647503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.655160] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f4b08 00:28:10.838 [2024-10-17 17:49:18.655911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:25300 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.655927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.663623] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166efae0 00:28:10.838 [2024-10-17 17:49:18.664389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:14095 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.664404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.672071] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e3d08 00:28:10.838 [2024-10-17 17:49:18.672818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23168 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.672834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.680534] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ed920 00:28:10.838 [2024-10-17 17:49:18.681286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:22978 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.681302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.688963] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ec840 00:28:10.838 [2024-10-17 17:49:18.689708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:3390 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.689724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.697377] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f6020 00:28:10.838 [2024-10-17 17:49:18.698127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:20840 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.698142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.705987] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e8088 00:28:10.838 [2024-10-17 17:49:18.706749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:8731 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.706764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.714456] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e6fa8 00:28:10.838 [2024-10-17 17:49:18.715206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:18836 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.715221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.722924] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166eb328 00:28:10.838 [2024-10-17 17:49:18.723686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:21064 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.723704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.731365] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f9b30 00:28:10.838 [2024-10-17 17:49:18.732104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:13847 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.732120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.739805] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f8a50 00:28:10.838 [2024-10-17 17:49:18.740556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:19479 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.740572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:10.838 [2024-10-17 17:49:18.748264] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166f7970 00:28:10.838 [2024-10-17 17:49:18.749019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:8571 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:10.838 [2024-10-17 17:49:18.749036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:11.099 [2024-10-17 17:49:18.756728] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166de8a8 00:28:11.099 [2024-10-17 17:49:18.757490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:17634 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.099 [2024-10-17 17:49:18.757506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.766931] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e23b8 00:28:11.100 [2024-10-17 17:49:18.768121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:11881 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.768136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.775505] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166dfdc0 00:28:11.100 [2024-10-17 17:49:18.776707] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:25085 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.776722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.783939] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166ebfd0 00:28:11.100 [2024-10-17 17:49:18.785094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:11294 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.785110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.792364] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166e23b8 00:28:11.100 [2024-10-17 17:49:18.793555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:6313 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.793571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.799728] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fd208 00:28:11.100 [2024-10-17 17:49:18.800615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:23002 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.800633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.808411] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.808654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:10498 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.808668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.817098] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.817302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:10575 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.817317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.825808] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.826054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:12271 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.826069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.834514] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.834761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:12307 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.834777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.843270] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.843529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6160 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.843544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.851996] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.852195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:12304 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.852210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.860700] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.860918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:15794 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.860933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.869425] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.869672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:14065 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.869687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.878137] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.878405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:21949 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.878420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.886828] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.887063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:10390 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.887078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.895599] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.895811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:15266 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.895826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.904291] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.904543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:3200 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.904558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.913037] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.913252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:19024 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.913267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.921718] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.921830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:11845 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.921845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.930380] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.930657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:23583 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.930673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:74 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 [2024-10-17 17:49:18.939117] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb200) with pdu=0x2000166fbcf0 00:28:11.100 [2024-10-17 17:49:18.939217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:15611 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:11.100 [2024-10-17 17:49:18.939231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:28:11.100 29986.50 IOPS, 117.13 MiB/s 00:28:11.100 Latency(us) 00:28:11.100 [2024-10-17T15:49:19.019Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:11.100 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:28:11.100 nvme0n1 : 2.00 29980.19 117.11 0.00 0.00 4262.65 2198.19 15728.64 00:28:11.100 [2024-10-17T15:49:19.019Z] =================================================================================================================== 00:28:11.100 [2024-10-17T15:49:19.019Z] Total : 29980.19 117.11 0.00 0.00 4262.65 2198.19 15728.64 00:28:11.100 { 00:28:11.100 "results": [ 00:28:11.100 { 00:28:11.100 "job": "nvme0n1", 00:28:11.100 "core_mask": "0x2", 00:28:11.100 "workload": "randwrite", 00:28:11.100 "status": "finished", 00:28:11.100 "queue_depth": 128, 00:28:11.100 "io_size": 4096, 00:28:11.100 "runtime": 2.004157, 00:28:11.100 "iops": 29980.186183018595, 00:28:11.100 "mibps": 117.11010227741639, 00:28:11.100 "io_failed": 0, 00:28:11.100 "io_timeout": 0, 00:28:11.100 "avg_latency_us": 4262.650993758841, 00:28:11.100 "min_latency_us": 2198.1866666666665, 00:28:11.100 "max_latency_us": 15728.64 00:28:11.100 } 00:28:11.100 ], 00:28:11.100 "core_count": 1 00:28:11.100 } 00:28:11.100 17:49:18 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:28:11.100 17:49:18 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:28:11.100 17:49:18 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:28:11.100 | .driver_specific 00:28:11.100 | .nvme_error 00:28:11.100 | .status_code 00:28:11.100 | .command_transient_transport_error' 00:28:11.100 17:49:18 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 235 > 0 )) 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 220305 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 220305 ']' 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 220305 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 220305 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 220305' 00:28:11.361 killing process with pid 220305 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 220305 00:28:11.361 Received shutdown signal, test time was about 2.000000 seconds 00:28:11.361 00:28:11.361 Latency(us) 00:28:11.361 [2024-10-17T15:49:19.280Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:11.361 [2024-10-17T15:49:19.280Z] =================================================================================================================== 00:28:11.361 [2024-10-17T15:49:19.280Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:11.361 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 220305 00:28:11.621 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@115 -- # run_bperf_err randwrite 131072 16 00:28:11.621 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:28:11.621 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:28:11.621 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:28:11.621 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=221007 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 221007 /var/tmp/bperf.sock 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 221007 ']' 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:28:11.622 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:11.622 17:49:19 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:11.622 [2024-10-17 17:49:19.360845] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:11.622 [2024-10-17 17:49:19.360902] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid221007 ] 00:28:11.622 I/O size of 131072 is greater than zero copy threshold (65536). 00:28:11.622 Zero copy mechanism will not be used. 00:28:11.622 [2024-10-17 17:49:19.437467] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:11.622 [2024-10-17 17:49:19.466648] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:12.563 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:28:12.822 nvme0n1 00:28:12.822 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:28:12.822 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:12.822 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:12.822 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:12.822 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:28:12.822 17:49:20 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:28:12.822 I/O size of 131072 is greater than zero copy threshold (65536). 00:28:12.822 Zero copy mechanism will not be used. 00:28:12.822 Running I/O for 2 seconds... 00:28:12.822 [2024-10-17 17:49:20.719168] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:12.822 [2024-10-17 17:49:20.719391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:12.822 [2024-10-17 17:49:20.719421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:12.822 [2024-10-17 17:49:20.728479] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:12.822 [2024-10-17 17:49:20.728544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:12.822 [2024-10-17 17:49:20.728563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:12.822 [2024-10-17 17:49:20.734724] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:12.822 [2024-10-17 17:49:20.734960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:12.822 [2024-10-17 17:49:20.734978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:12.822 [2024-10-17 17:49:20.741633] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.083 [2024-10-17 17:49:20.741688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.083 [2024-10-17 17:49:20.741711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.083 [2024-10-17 17:49:20.749054] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.083 [2024-10-17 17:49:20.749405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.083 [2024-10-17 17:49:20.749424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.083 [2024-10-17 17:49:20.755856] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.083 [2024-10-17 17:49:20.756171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.083 [2024-10-17 17:49:20.756189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.764381] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.764748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.764767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.771324] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.771638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.771657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.777334] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.777651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.777672] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.785030] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.785355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.785372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.790379] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.790568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.790584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.797849] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.798155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.798172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.807816] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.808185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.808202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.815706] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.815898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.815914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.821141] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.821331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.821347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.827261] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.827651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.827668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.835171] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.835487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.835505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.844148] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.844466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.844483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.852213] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.852528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.852545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.859911] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.860107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.860124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.867331] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.867520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.867536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.875584] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.875832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.875849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.882547] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.882741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.882757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.891196] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.891395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.891411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.895413] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.895604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.895620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.903498] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.903704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.903720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.914809] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.915118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.915136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.923600] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.923923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.923941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.935362] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.935695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.935713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.947335] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.947729] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.947746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.959574] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.959778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.959794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.971213] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.971662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.971679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.982421] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.982687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.982707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.084 [2024-10-17 17:49:20.993886] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.084 [2024-10-17 17:49:20.994129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.084 [2024-10-17 17:49:20.994146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.005951] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.006236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.006256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.017241] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.017561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.017578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.028735] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.029008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.029024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.040132] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.040487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.040504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.052101] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.052404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.052421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.064208] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.064634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.064652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.076029] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.076344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.076362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.087934] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.088252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.088270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.099833] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.100182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.100199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.111434] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.111758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.111776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.123877] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.124202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.124219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.135778] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.136267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.136283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.146564] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.146875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.146893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.157705] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.158051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.158069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.168134] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.168429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.168446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.178220] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.178639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.178657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.190042] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.190265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.190282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.200672] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.201006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.201027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.211749] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.212194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.212212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.218637] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.218844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.218861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.226418] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.226741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.226758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.233426] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.233861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.345 [2024-10-17 17:49:21.233879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.345 [2024-10-17 17:49:21.242607] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.345 [2024-10-17 17:49:21.242941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.346 [2024-10-17 17:49:21.242956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.346 [2024-10-17 17:49:21.250515] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.346 [2024-10-17 17:49:21.250832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.346 [2024-10-17 17:49:21.250850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.346 [2024-10-17 17:49:21.256901] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.346 [2024-10-17 17:49:21.257244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.346 [2024-10-17 17:49:21.257261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.265127] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.265491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.265508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.273545] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.273935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.273952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.280352] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.280668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.280685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.287211] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.287401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.287417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.292871] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.293061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.293077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.298097] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.298285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.298301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.307118] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.307545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.307562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.315309] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.315500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.315517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.321018] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.321427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.321445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.327980] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.328314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.328331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.337591] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.337892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.337910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.345700] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.346046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.607 [2024-10-17 17:49:21.346063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.607 [2024-10-17 17:49:21.355448] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.607 [2024-10-17 17:49:21.355763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.355780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.361272] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.361462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.361478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.367837] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.368028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.368045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.376254] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.376577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.376595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.380637] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.380929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.380947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.388162] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.388463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.388480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.395229] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.395615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.395636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.403258] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.403553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.403570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.410075] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.410492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.410509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.418453] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.418874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.418892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.426887] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.427086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.427103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.435063] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.435253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.435269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.443444] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.443887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.443905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.452271] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.452592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.452609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.461204] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.461612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.461630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.468407] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.468620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.468635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.478604] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.478911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.478928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.487003] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.487302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.487319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.495853] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.496165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.496182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.503217] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.503406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.503422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.511650] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.511993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.512011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.608 [2024-10-17 17:49:21.520397] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.608 [2024-10-17 17:49:21.520658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.608 [2024-10-17 17:49:21.520673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.529994] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.530193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.530210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.538262] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.538577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.538594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.545339] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.545545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.545561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.553331] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.553648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.553665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.562223] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.562568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.562585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.570922] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.571288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.571305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.579235] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.579551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.579569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.587214] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.587415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.587431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.597804] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.598240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.598256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.608891] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.609263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.609280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.616187] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.616247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.616265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.625035] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.625318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.625335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.632011] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.632341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.632358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.641289] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.641598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.641615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.648766] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.649136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.649153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.656645] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.657037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.657054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.664588] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.664903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.664920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.673443] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.673506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.673521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.682167] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.682467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.682482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.689112] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.689168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.689183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.694552] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.694609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.694624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.698396] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.698469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.698484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.870 [2024-10-17 17:49:21.702793] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.702845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.702860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.870 3597.00 IOPS, 449.62 MiB/s [2024-10-17T15:49:21.789Z] [2024-10-17 17:49:21.711912] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.870 [2024-10-17 17:49:21.712151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.870 [2024-10-17 17:49:21.712167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.720364] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.720594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.720610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.729211] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.729434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.729450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.735084] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.735139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.735155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.738315] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.738371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.738390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.741443] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.741496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.741512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.745967] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.746034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.746049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.749473] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.749531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.749546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.754056] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.754324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.754340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.758411] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.758467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.758482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.764769] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.765045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.765060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.772549] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.772814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.772829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:13.871 [2024-10-17 17:49:21.781426] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:13.871 [2024-10-17 17:49:21.781740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:13.871 [2024-10-17 17:49:21.781757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.132 [2024-10-17 17:49:21.790238] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.132 [2024-10-17 17:49:21.790536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.132 [2024-10-17 17:49:21.790552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.132 [2024-10-17 17:49:21.799812] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.132 [2024-10-17 17:49:21.800067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.800083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.810097] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.810356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.810371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.820739] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.821006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.821021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.830871] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.831151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.831166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.841161] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.841415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.841430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.851286] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.851493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.851508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.862133] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.862375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.862390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.872529] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.872760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.872776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.882650] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.882911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.882927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.892988] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.893235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.893250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.903174] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.903424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.903439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.912653] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.912882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.912897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.923302] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.923605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.923621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.931905] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.932099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.932114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.942137] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.942421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.942438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.951326] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.951537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.951553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.961527] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.961791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.961809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.969653] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.969926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.969942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.978147] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.978466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.978482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.987768] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.988139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.988155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:21.997361] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:21.997584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:21.997599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:22.006704] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:22.006767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:22.006782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:22.015365] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:22.015558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:22.015573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:22.026641] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:22.026932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:22.026948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:22.038026] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:22.038079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:22.038094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.133 [2024-10-17 17:49:22.049880] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.133 [2024-10-17 17:49:22.050210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.133 [2024-10-17 17:49:22.050226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.060222] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.060340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.060355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.068054] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.068103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.068118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.076119] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.076162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.076176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.079470] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.079514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.079529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.082902] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.082947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.082962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.086155] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.086203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.086218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.091962] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.092010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.092025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.095380] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.395 [2024-10-17 17:49:22.095425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.395 [2024-10-17 17:49:22.095440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.395 [2024-10-17 17:49:22.100965] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.101245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.101260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.106586] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.106639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.106654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.115142] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.115399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.115414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.120232] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.120476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.120491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.126302] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.126361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.126376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.132213] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.132267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.132282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.135964] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.136009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.136024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.139354] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.139396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.139411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.142718] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.142776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.142794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.149800] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.149844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.149859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.153151] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.153194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.153209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.156507] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.156584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.156599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.159837] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.159892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.159907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.163300] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.163343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.163358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.170241] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.170468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.170483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.176067] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.176304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.176319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.182606] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.182684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.182706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.190203] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.190271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.190286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.193569] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.193645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.193660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.197019] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.197090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.197106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.200536] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.200610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.200625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.204221] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.204287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.204302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.211030] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.211328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.211344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.220757] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.221092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.221108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.230527] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.230808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.230823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.241139] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.241380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.241395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.252064] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.252175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.252190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.396 [2024-10-17 17:49:22.262165] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.396 [2024-10-17 17:49:22.262416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.396 [2024-10-17 17:49:22.262438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.397 [2024-10-17 17:49:22.272544] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.397 [2024-10-17 17:49:22.272844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.397 [2024-10-17 17:49:22.272866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.397 [2024-10-17 17:49:22.283284] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.397 [2024-10-17 17:49:22.283572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.397 [2024-10-17 17:49:22.283588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.397 [2024-10-17 17:49:22.293942] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.397 [2024-10-17 17:49:22.294197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.397 [2024-10-17 17:49:22.294212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.397 [2024-10-17 17:49:22.304297] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.397 [2024-10-17 17:49:22.304402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.397 [2024-10-17 17:49:22.304417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.314443] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.314725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.314740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.324666] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.324929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.324945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.334801] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.335059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.335077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.344222] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.344519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.344534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.354625] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.354750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.354765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.364260] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.364527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.364551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.373616] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.373941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.373957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.383711] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.383946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.383962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.393981] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.394166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.394182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.399433] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.399702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.399717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.407737] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.408056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.408073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.412592] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.412663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.412678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.417496] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.417577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.417593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.422943] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.423260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.423277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.428014] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.428272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.428287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.431247] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.431313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.658 [2024-10-17 17:49:22.431329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.658 [2024-10-17 17:49:22.434130] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.658 [2024-10-17 17:49:22.434210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.434225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.437100] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.437175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.437190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.440041] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.440109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.440124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.442849] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.442922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.442937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.445790] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.445925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.445941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.449327] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.449599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.449616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.459133] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.459478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.459495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.469546] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.469820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.469836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.480556] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.480781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.480797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.491042] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.491309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.491326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.501369] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.501683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.501703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.511869] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.512098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.512114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.522980] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.523245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.523266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.533193] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.533448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.533463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.544283] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.544371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.544386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.554716] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.555064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.555080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.565798] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.566056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.566073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.659 [2024-10-17 17:49:22.575931] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.659 [2024-10-17 17:49:22.576250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.659 [2024-10-17 17:49:22.576266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.920 [2024-10-17 17:49:22.586446] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.920 [2024-10-17 17:49:22.586704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.586720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.597065] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.597274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.597290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.606677] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.606968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.606984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.617469] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.617709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.617724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.628266] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.628527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.628542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.638514] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.638603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.638618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.649022] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.649231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.649246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.658538] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.658854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.658870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.668533] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.668793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.668808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.678285] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.678365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.678380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.687626] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.687769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.687784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.696893] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.697112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.697128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:28:14.921 [2024-10-17 17:49:22.701847] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1edb540) with pdu=0x2000166fef90 00:28:14.921 [2024-10-17 17:49:22.701907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:14.921 [2024-10-17 17:49:22.701922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:28:14.921 3774.00 IOPS, 471.75 MiB/s 00:28:14.921 Latency(us) 00:28:14.921 [2024-10-17T15:49:22.840Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:14.921 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:28:14.921 nvme0n1 : 2.00 3775.97 472.00 0.00 0.00 4232.95 1331.20 15510.19 00:28:14.921 [2024-10-17T15:49:22.840Z] =================================================================================================================== 00:28:14.921 [2024-10-17T15:49:22.840Z] Total : 3775.97 472.00 0.00 0.00 4232.95 1331.20 15510.19 00:28:14.921 { 00:28:14.921 "results": [ 00:28:14.921 { 00:28:14.921 "job": "nvme0n1", 00:28:14.921 "core_mask": "0x2", 00:28:14.921 "workload": "randwrite", 00:28:14.921 "status": "finished", 00:28:14.921 "queue_depth": 16, 00:28:14.921 "io_size": 131072, 00:28:14.921 "runtime": 2.003193, 00:28:14.921 "iops": 3775.971661242826, 00:28:14.921 "mibps": 471.99645765535325, 00:28:14.921 "io_failed": 0, 00:28:14.921 "io_timeout": 0, 00:28:14.921 "avg_latency_us": 4232.952554204126, 00:28:14.921 "min_latency_us": 1331.2, 00:28:14.921 "max_latency_us": 15510.186666666666 00:28:14.921 } 00:28:14.921 ], 00:28:14.921 "core_count": 1 00:28:14.921 } 00:28:14.921 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:28:14.921 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:28:14.921 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:28:14.921 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:28:14.921 | .driver_specific 00:28:14.921 | .nvme_error 00:28:14.921 | .status_code 00:28:14.921 | .command_transient_transport_error' 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 243 > 0 )) 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 221007 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 221007 ']' 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 221007 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 221007 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 221007' 00:28:15.182 killing process with pid 221007 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 221007 00:28:15.182 Received shutdown signal, test time was about 2.000000 seconds 00:28:15.182 00:28:15.182 Latency(us) 00:28:15.182 [2024-10-17T15:49:23.101Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:15.182 [2024-10-17T15:49:23.101Z] =================================================================================================================== 00:28:15.182 [2024-10-17T15:49:23.101Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:15.182 17:49:22 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 221007 00:28:15.182 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@116 -- # killprocess 218586 00:28:15.182 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 218586 ']' 00:28:15.182 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 218586 00:28:15.182 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:28:15.182 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:15.182 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 218586 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 218586' 00:28:15.443 killing process with pid 218586 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 218586 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 218586 00:28:15.443 00:28:15.443 real 0m16.486s 00:28:15.443 user 0m32.615s 00:28:15.443 sys 0m3.635s 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:28:15.443 ************************************ 00:28:15.443 END TEST nvmf_digest_error 00:28:15.443 ************************************ 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@150 -- # nvmftestfini 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@514 -- # nvmfcleanup 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@121 -- # sync 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@124 -- # set +e 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@125 -- # for i in {1..20} 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:28:15.443 rmmod nvme_tcp 00:28:15.443 rmmod nvme_fabrics 00:28:15.443 rmmod nvme_keyring 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@128 -- # set -e 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@129 -- # return 0 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@515 -- # '[' -n 218586 ']' 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@516 -- # killprocess 218586 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@950 -- # '[' -z 218586 ']' 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@954 -- # kill -0 218586 00:28:15.443 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (218586) - No such process 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@977 -- # echo 'Process with pid 218586 is not found' 00:28:15.443 Process with pid 218586 is not found 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:28:15.443 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@297 -- # iptr 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@789 -- # iptables-save 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@789 -- # iptables-restore 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@302 -- # remove_spdk_ns 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:15.704 17:49:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:17.616 17:49:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:28:17.617 00:28:17.617 real 0m43.496s 00:28:17.617 user 1m8.265s 00:28:17.617 sys 0m13.199s 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:28:17.617 ************************************ 00:28:17.617 END TEST nvmf_digest 00:28:17.617 ************************************ 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@36 -- # [[ 0 -eq 1 ]] 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@41 -- # [[ 0 -eq 1 ]] 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@46 -- # [[ phy == phy ]] 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@47 -- # run_test nvmf_bdevperf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:28:17.617 ************************************ 00:28:17.617 START TEST nvmf_bdevperf 00:28:17.617 ************************************ 00:28:17.617 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:28:17.878 * Looking for test storage... 00:28:17.878 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1691 -- # lcov --version 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # IFS=.-: 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # read -ra ver1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # IFS=.-: 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # read -ra ver2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@338 -- # local 'op=<' 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@340 -- # ver1_l=2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@341 -- # ver2_l=1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@344 -- # case "$op" in 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@345 -- # : 1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v = 0 )) 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # decimal 1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # ver1[v]=1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # decimal 2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # ver2[v]=2 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # return 0 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:28:17.878 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:17.878 --rc genhtml_branch_coverage=1 00:28:17.878 --rc genhtml_function_coverage=1 00:28:17.878 --rc genhtml_legend=1 00:28:17.878 --rc geninfo_all_blocks=1 00:28:17.878 --rc geninfo_unexecuted_blocks=1 00:28:17.878 00:28:17.878 ' 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:28:17.878 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:17.878 --rc genhtml_branch_coverage=1 00:28:17.878 --rc genhtml_function_coverage=1 00:28:17.878 --rc genhtml_legend=1 00:28:17.878 --rc geninfo_all_blocks=1 00:28:17.878 --rc geninfo_unexecuted_blocks=1 00:28:17.878 00:28:17.878 ' 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:28:17.878 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:17.878 --rc genhtml_branch_coverage=1 00:28:17.878 --rc genhtml_function_coverage=1 00:28:17.878 --rc genhtml_legend=1 00:28:17.878 --rc geninfo_all_blocks=1 00:28:17.878 --rc geninfo_unexecuted_blocks=1 00:28:17.878 00:28:17.878 ' 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:28:17.878 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:17.878 --rc genhtml_branch_coverage=1 00:28:17.878 --rc genhtml_function_coverage=1 00:28:17.878 --rc genhtml_legend=1 00:28:17.878 --rc geninfo_all_blocks=1 00:28:17.878 --rc geninfo_unexecuted_blocks=1 00:28:17.878 00:28:17.878 ' 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # uname -s 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:17.878 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@15 -- # shopt -s extglob 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@5 -- # export PATH 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@51 -- # : 0 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:28:17.879 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@11 -- # MALLOC_BDEV_SIZE=64 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@24 -- # nvmftestinit 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@474 -- # prepare_net_devs 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@436 -- # local -g is_hw=no 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@438 -- # remove_spdk_ns 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@309 -- # xtrace_disable 00:28:17.879 17:49:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # pci_devs=() 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # local -a pci_devs 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # pci_drivers=() 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # net_devs=() 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # local -ga net_devs 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # e810=() 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # local -ga e810 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # x722=() 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # local -ga x722 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # mlx=() 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # local -ga mlx 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:26.021 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:28:26.021 Found 0000:31:00.0 (0x8086 - 0x159b) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:28:26.022 Found 0000:31:00.1 (0x8086 - 0x159b) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:28:26.022 Found net devices under 0000:31:00.0: cvl_0_0 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:28:26.022 Found net devices under 0000:31:00.1: cvl_0_1 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # is_hw=yes 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:28:26.022 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:28:26.022 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.646 ms 00:28:26.022 00:28:26.022 --- 10.0.0.2 ping statistics --- 00:28:26.022 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:26.022 rtt min/avg/max/mdev = 0.646/0.646/0.646/0.000 ms 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:28:26.022 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:28:26.022 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.290 ms 00:28:26.022 00:28:26.022 --- 10.0.0.1 ping statistics --- 00:28:26.022 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:26.022 rtt min/avg/max/mdev = 0.290/0.290/0.290/0.000 ms 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@448 -- # return 0 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@25 -- # tgt_init 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@724 -- # xtrace_disable 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # nvmfpid=226047 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # waitforlisten 226047 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@831 -- # '[' -z 226047 ']' 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:26.022 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:26.022 17:49:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.022 [2024-10-17 17:49:33.459315] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:26.022 [2024-10-17 17:49:33.459376] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:26.022 [2024-10-17 17:49:33.551488] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:28:26.022 [2024-10-17 17:49:33.603992] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:28:26.022 [2024-10-17 17:49:33.604040] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:28:26.022 [2024-10-17 17:49:33.604049] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:28:26.022 [2024-10-17 17:49:33.604056] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:28:26.022 [2024-10-17 17:49:33.604063] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:28:26.022 [2024-10-17 17:49:33.605975] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:28:26.022 [2024-10-17 17:49:33.606135] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:26.022 [2024-10-17 17:49:33.606134] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # return 0 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@730 -- # xtrace_disable 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.595 [2024-10-17 17:49:34.325600] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.595 Malloc0 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.595 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:26.596 [2024-10-17 17:49:34.397378] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 128 -o 4096 -w verify -t 1 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # gen_nvmf_target_json 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # config=() 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # local subsystem config 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:28:26.596 { 00:28:26.596 "params": { 00:28:26.596 "name": "Nvme$subsystem", 00:28:26.596 "trtype": "$TEST_TRANSPORT", 00:28:26.596 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:26.596 "adrfam": "ipv4", 00:28:26.596 "trsvcid": "$NVMF_PORT", 00:28:26.596 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:26.596 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:26.596 "hdgst": ${hdgst:-false}, 00:28:26.596 "ddgst": ${ddgst:-false} 00:28:26.596 }, 00:28:26.596 "method": "bdev_nvme_attach_controller" 00:28:26.596 } 00:28:26.596 EOF 00:28:26.596 )") 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # cat 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # jq . 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@583 -- # IFS=, 00:28:26.596 17:49:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:28:26.596 "params": { 00:28:26.596 "name": "Nvme1", 00:28:26.596 "trtype": "tcp", 00:28:26.596 "traddr": "10.0.0.2", 00:28:26.596 "adrfam": "ipv4", 00:28:26.596 "trsvcid": "4420", 00:28:26.596 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:28:26.596 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:28:26.596 "hdgst": false, 00:28:26.596 "ddgst": false 00:28:26.596 }, 00:28:26.596 "method": "bdev_nvme_attach_controller" 00:28:26.596 }' 00:28:26.596 [2024-10-17 17:49:34.463827] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:26.596 [2024-10-17 17:49:34.463890] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid226331 ] 00:28:26.856 [2024-10-17 17:49:34.545212] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:26.856 [2024-10-17 17:49:34.597927] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:26.856 Running I/O for 1 seconds... 00:28:28.242 8594.00 IOPS, 33.57 MiB/s 00:28:28.242 Latency(us) 00:28:28.242 [2024-10-17T15:49:36.161Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:28.242 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:28:28.242 Verification LBA range: start 0x0 length 0x4000 00:28:28.242 Nvme1n1 : 1.01 8631.98 33.72 0.00 0.00 14766.10 3317.76 14308.69 00:28:28.242 [2024-10-17T15:49:36.161Z] =================================================================================================================== 00:28:28.242 [2024-10-17T15:49:36.161Z] Total : 8631.98 33.72 0.00 0.00 14766.10 3317.76 14308.69 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@30 -- # bdevperfpid=226550 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@32 -- # sleep 3 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -q 128 -o 4096 -w verify -t 15 -f 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # gen_nvmf_target_json 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # config=() 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # local subsystem config 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:28:28.242 { 00:28:28.242 "params": { 00:28:28.242 "name": "Nvme$subsystem", 00:28:28.242 "trtype": "$TEST_TRANSPORT", 00:28:28.242 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:28.242 "adrfam": "ipv4", 00:28:28.242 "trsvcid": "$NVMF_PORT", 00:28:28.242 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:28.242 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:28.242 "hdgst": ${hdgst:-false}, 00:28:28.242 "ddgst": ${ddgst:-false} 00:28:28.242 }, 00:28:28.242 "method": "bdev_nvme_attach_controller" 00:28:28.242 } 00:28:28.242 EOF 00:28:28.242 )") 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # cat 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # jq . 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@583 -- # IFS=, 00:28:28.242 17:49:35 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:28:28.242 "params": { 00:28:28.242 "name": "Nvme1", 00:28:28.242 "trtype": "tcp", 00:28:28.242 "traddr": "10.0.0.2", 00:28:28.242 "adrfam": "ipv4", 00:28:28.242 "trsvcid": "4420", 00:28:28.242 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:28:28.242 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:28:28.242 "hdgst": false, 00:28:28.242 "ddgst": false 00:28:28.242 }, 00:28:28.242 "method": "bdev_nvme_attach_controller" 00:28:28.242 }' 00:28:28.242 [2024-10-17 17:49:35.975003] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:28.242 [2024-10-17 17:49:35.975059] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid226550 ] 00:28:28.242 [2024-10-17 17:49:36.053709] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:28.242 [2024-10-17 17:49:36.088734] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:28.503 Running I/O for 15 seconds... 00:28:30.829 9047.00 IOPS, 35.34 MiB/s [2024-10-17T15:49:39.011Z] 10163.50 IOPS, 39.70 MiB/s [2024-10-17T15:49:39.011Z] 17:49:38 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@33 -- # kill -9 226047 00:28:31.092 17:49:38 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@35 -- # sleep 3 00:28:31.092 [2024-10-17 17:49:38.940486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:78400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:78408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:78416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:78424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:78432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:78440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:78448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:78456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:78464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:78472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:78480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:78488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:78496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:78504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:78512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:78520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:78528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:78536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.940991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:78544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.940999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.941012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:78552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.941021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.092 [2024-10-17 17:49:38.941033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:78560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.092 [2024-10-17 17:49:38.941041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:78568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:78576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:78584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:78592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:78600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:78608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:78616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:78624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:78632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:78640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:78648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:78656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:78664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:78672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:78680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:78688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:78696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:78704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:78712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:78720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:78728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:78736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941413] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:78744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:78752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:78760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:78768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:78776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:78784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:78792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:78800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:78808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:78816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:78824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:78832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:78840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:78848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:78856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:78864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:78872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:78880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.093 [2024-10-17 17:49:38.941728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:78888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.093 [2024-10-17 17:49:38.941735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:78896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:78904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:78912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:78920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:78928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:78936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:78944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:78952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:78960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:78968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:78976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:78984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:78992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:79000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:79008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.941989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.941998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:79016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:79024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:79032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:79040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:79048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:79056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:79064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:79072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:79080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:79088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:79096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:79104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:79368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:31.094 [2024-10-17 17:49:38.942206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:79376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:31.094 [2024-10-17 17:49:38.942222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:79384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:31.094 [2024-10-17 17:49:38.942239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:79392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:31.094 [2024-10-17 17:49:38.942260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:79400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:31.094 [2024-10-17 17:49:38.942277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:79408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:31.094 [2024-10-17 17:49:38.942293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:79416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:31.094 [2024-10-17 17:49:38.942310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:79112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:79120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:79128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:79136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:79144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:79152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:79160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.094 [2024-10-17 17:49:38.942439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:79168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.094 [2024-10-17 17:49:38.942446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:79176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:79184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:79192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:79200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:79208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:79216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:79224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:79232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:79240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:79248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:79256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:79264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:79272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:79280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:79288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:79296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:79304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:79312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:79320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:79328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:79336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:79344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:79352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:31.095 [2024-10-17 17:49:38.942836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942846] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60a6a0 is same with the state(6) to be set 00:28:31.095 [2024-10-17 17:49:38.942856] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:31.095 [2024-10-17 17:49:38.942862] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:31.095 [2024-10-17 17:49:38.942869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:79360 len:8 PRP1 0x0 PRP2 0x0 00:28:31.095 [2024-10-17 17:49:38.942877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:31.095 [2024-10-17 17:49:38.942915] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x60a6a0 was disconnected and freed. reset controller. 00:28:31.095 [2024-10-17 17:49:38.946507] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.095 [2024-10-17 17:49:38.946559] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.095 [2024-10-17 17:49:38.947286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.095 [2024-10-17 17:49:38.947303] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.095 [2024-10-17 17:49:38.947311] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.095 [2024-10-17 17:49:38.947528] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.095 [2024-10-17 17:49:38.947751] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.095 [2024-10-17 17:49:38.947760] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.095 [2024-10-17 17:49:38.947769] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.095 [2024-10-17 17:49:38.951258] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.095 [2024-10-17 17:49:38.960511] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.095 [2024-10-17 17:49:38.961007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.095 [2024-10-17 17:49:38.961046] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.095 [2024-10-17 17:49:38.961057] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.095 [2024-10-17 17:49:38.961294] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.095 [2024-10-17 17:49:38.961514] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.095 [2024-10-17 17:49:38.961522] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.095 [2024-10-17 17:49:38.961530] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.095 [2024-10-17 17:49:38.965040] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.095 [2024-10-17 17:49:38.974297] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.095 [2024-10-17 17:49:38.974830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.095 [2024-10-17 17:49:38.974869] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.095 [2024-10-17 17:49:38.974882] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.095 [2024-10-17 17:49:38.975120] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.095 [2024-10-17 17:49:38.975345] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.095 [2024-10-17 17:49:38.975355] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.095 [2024-10-17 17:49:38.975363] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.095 [2024-10-17 17:49:38.978864] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.095 [2024-10-17 17:49:38.988128] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.095 [2024-10-17 17:49:38.988815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.095 [2024-10-17 17:49:38.988855] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.095 [2024-10-17 17:49:38.988868] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.095 [2024-10-17 17:49:38.989105] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.095 [2024-10-17 17:49:38.989325] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.095 [2024-10-17 17:49:38.989335] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.095 [2024-10-17 17:49:38.989343] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.095 [2024-10-17 17:49:38.992844] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.096 [2024-10-17 17:49:39.001904] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.096 [2024-10-17 17:49:39.002558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.096 [2024-10-17 17:49:39.002598] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.096 [2024-10-17 17:49:39.002609] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.096 [2024-10-17 17:49:39.002856] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.096 [2024-10-17 17:49:39.003077] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.096 [2024-10-17 17:49:39.003085] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.096 [2024-10-17 17:49:39.003094] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.096 [2024-10-17 17:49:39.006589] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.358 [2024-10-17 17:49:39.015653] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.358 [2024-10-17 17:49:39.016314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.016355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.016367] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.016604] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.016837] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.016847] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.016855] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.020354] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.029401] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.030067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.030109] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.030120] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.030358] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.030578] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.030587] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.030595] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.034101] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.043157] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.043618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.043641] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.043650] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.043875] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.044093] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.044102] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.044109] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.047603] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.057065] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.057606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.057623] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.057631] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.057854] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.058071] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.058085] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.058093] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.061619] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.070881] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.071524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.071571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.071588] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.071841] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.072063] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.072072] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.072080] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.075585] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.084667] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.085350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.085403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.085415] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.085660] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.085891] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.085901] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.085909] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.089413] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.098493] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.099191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.099246] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.099258] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.099506] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.099739] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.099749] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.099758] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.103265] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.112344] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.113032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.113091] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.113104] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.113354] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.113592] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.113611] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.113619] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.117164] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.126309] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.127032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.127094] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.127107] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.127359] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.127583] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.127592] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.127601] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.131129] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.140213] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.140799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.140827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.140836] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.141056] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.141276] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.141285] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.141293] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.144809] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.359 [2024-10-17 17:49:39.154087] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.359 [2024-10-17 17:49:39.154655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.359 [2024-10-17 17:49:39.154677] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.359 [2024-10-17 17:49:39.154685] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.359 [2024-10-17 17:49:39.154911] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.359 [2024-10-17 17:49:39.155131] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.359 [2024-10-17 17:49:39.155140] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.359 [2024-10-17 17:49:39.155148] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.359 [2024-10-17 17:49:39.158649] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.167933] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.168591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.168653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.168666] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.168929] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.169154] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.169164] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.169172] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.172696] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.181804] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.182390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.182423] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.182432] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.182651] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.182880] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.182889] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.182897] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.186403] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.195703] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.196340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.196363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.196373] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.196592] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.196829] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.196840] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.196847] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.200354] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.209451] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.210078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.210102] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.210112] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.210340] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.210559] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.210569] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.210576] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.214109] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.223389] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.223959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.223984] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.223994] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.224214] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.224434] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.224443] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.224450] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.227962] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.237238] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.237989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.238054] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.238066] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.238319] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.238542] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.238552] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.238560] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.242087] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.251164] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.251753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.251782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.251791] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.252012] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.252230] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.252240] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.252255] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.255774] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.360 [2024-10-17 17:49:39.265054] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.360 [2024-10-17 17:49:39.265617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.360 [2024-10-17 17:49:39.265639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.360 [2024-10-17 17:49:39.265647] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.360 [2024-10-17 17:49:39.265873] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.360 [2024-10-17 17:49:39.266092] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.360 [2024-10-17 17:49:39.266102] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.360 [2024-10-17 17:49:39.266110] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.360 [2024-10-17 17:49:39.269654] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.623 [2024-10-17 17:49:39.278941] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.623 [2024-10-17 17:49:39.279548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.623 [2024-10-17 17:49:39.279570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.623 [2024-10-17 17:49:39.279579] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.623 [2024-10-17 17:49:39.279803] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.623 [2024-10-17 17:49:39.280022] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.623 [2024-10-17 17:49:39.280031] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.623 [2024-10-17 17:49:39.280039] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.623 [2024-10-17 17:49:39.283564] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.623 [2024-10-17 17:49:39.292837] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.623 [2024-10-17 17:49:39.293532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.623 [2024-10-17 17:49:39.293597] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.623 [2024-10-17 17:49:39.293610] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.623 [2024-10-17 17:49:39.293876] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.623 [2024-10-17 17:49:39.294100] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.623 [2024-10-17 17:49:39.294110] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.623 [2024-10-17 17:49:39.294118] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.623 [2024-10-17 17:49:39.297654] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.623 [2024-10-17 17:49:39.306750] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.623 [2024-10-17 17:49:39.307374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.623 [2024-10-17 17:49:39.307410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.623 [2024-10-17 17:49:39.307420] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.623 [2024-10-17 17:49:39.307639] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.307868] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.307878] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.307886] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.311387] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.320681] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.321354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.321416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.321429] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.321681] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.321913] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.321924] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.321933] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.325455] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.334530] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.335243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.335306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.335319] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.335571] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.335808] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.335819] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.335828] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.339351] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.348422] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.349044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.349108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.349121] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.349373] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.349605] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.349615] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.349623] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.353150] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.362235] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.363025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.363088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.363101] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.363353] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.363576] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.363585] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.363594] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.367120] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.376002] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.376739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.376803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.376818] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.377072] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.377294] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.377305] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.377314] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.380862] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.389948] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.390590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.390618] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.390627] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.390855] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.391075] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.391085] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.391093] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.394613] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.403712] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.404382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.404445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.404459] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.404721] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.404945] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.404956] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.404964] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.410141] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 8728.00 IOPS, 34.09 MiB/s [2024-10-17T15:49:39.543Z] [2024-10-17 17:49:39.417613] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.418210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.418274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.418287] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.418540] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.418777] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.418787] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.418796] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.422313] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.431384] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.432031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.432094] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.432107] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.432359] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.432583] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.432592] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.432601] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.436126] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.445208] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.445822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.445886] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.624 [2024-10-17 17:49:39.445907] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.624 [2024-10-17 17:49:39.446161] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.624 [2024-10-17 17:49:39.446384] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.624 [2024-10-17 17:49:39.446394] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.624 [2024-10-17 17:49:39.446402] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.624 [2024-10-17 17:49:39.449936] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.624 [2024-10-17 17:49:39.459026] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.624 [2024-10-17 17:49:39.459606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.624 [2024-10-17 17:49:39.459634] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.625 [2024-10-17 17:49:39.459643] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.625 [2024-10-17 17:49:39.459872] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.625 [2024-10-17 17:49:39.460091] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.625 [2024-10-17 17:49:39.460100] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.625 [2024-10-17 17:49:39.460109] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.625 [2024-10-17 17:49:39.463881] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.625 [2024-10-17 17:49:39.472971] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.625 [2024-10-17 17:49:39.473568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.625 [2024-10-17 17:49:39.473594] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.625 [2024-10-17 17:49:39.473603] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.625 [2024-10-17 17:49:39.473830] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.625 [2024-10-17 17:49:39.474049] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.625 [2024-10-17 17:49:39.474059] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.625 [2024-10-17 17:49:39.474067] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.625 [2024-10-17 17:49:39.477616] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.625 [2024-10-17 17:49:39.486918] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.625 [2024-10-17 17:49:39.487594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.625 [2024-10-17 17:49:39.487658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.625 [2024-10-17 17:49:39.487672] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.625 [2024-10-17 17:49:39.487936] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.625 [2024-10-17 17:49:39.488160] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.625 [2024-10-17 17:49:39.488177] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.625 [2024-10-17 17:49:39.488186] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.625 [2024-10-17 17:49:39.491709] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.625 [2024-10-17 17:49:39.500799] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.625 [2024-10-17 17:49:39.501413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.625 [2024-10-17 17:49:39.501476] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.625 [2024-10-17 17:49:39.501489] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.625 [2024-10-17 17:49:39.501752] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.625 [2024-10-17 17:49:39.501977] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.625 [2024-10-17 17:49:39.501988] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.625 [2024-10-17 17:49:39.501997] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.625 [2024-10-17 17:49:39.505513] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.625 [2024-10-17 17:49:39.514586] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.625 [2024-10-17 17:49:39.515059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.625 [2024-10-17 17:49:39.515089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.625 [2024-10-17 17:49:39.515098] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.625 [2024-10-17 17:49:39.515316] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.625 [2024-10-17 17:49:39.515535] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.625 [2024-10-17 17:49:39.515545] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.625 [2024-10-17 17:49:39.515554] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.625 [2024-10-17 17:49:39.519074] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.625 [2024-10-17 17:49:39.528346] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.625 [2024-10-17 17:49:39.528840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.625 [2024-10-17 17:49:39.528863] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.625 [2024-10-17 17:49:39.528872] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.625 [2024-10-17 17:49:39.529091] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.625 [2024-10-17 17:49:39.529309] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.625 [2024-10-17 17:49:39.529318] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.625 [2024-10-17 17:49:39.529325] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.625 [2024-10-17 17:49:39.532838] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.888 [2024-10-17 17:49:39.542132] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.888 [2024-10-17 17:49:39.542637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.888 [2024-10-17 17:49:39.542658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.888 [2024-10-17 17:49:39.542667] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.888 [2024-10-17 17:49:39.542892] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.888 [2024-10-17 17:49:39.543109] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.888 [2024-10-17 17:49:39.543125] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.888 [2024-10-17 17:49:39.543133] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.888 [2024-10-17 17:49:39.546635] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.888 [2024-10-17 17:49:39.555909] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.888 [2024-10-17 17:49:39.556564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.888 [2024-10-17 17:49:39.556627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.888 [2024-10-17 17:49:39.556639] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.888 [2024-10-17 17:49:39.556904] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.888 [2024-10-17 17:49:39.557129] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.888 [2024-10-17 17:49:39.557139] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.888 [2024-10-17 17:49:39.557147] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.888 [2024-10-17 17:49:39.560666] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.888 [2024-10-17 17:49:39.569744] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.888 [2024-10-17 17:49:39.570439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.888 [2024-10-17 17:49:39.570503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.888 [2024-10-17 17:49:39.570515] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.888 [2024-10-17 17:49:39.570779] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.888 [2024-10-17 17:49:39.571004] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.888 [2024-10-17 17:49:39.571015] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.888 [2024-10-17 17:49:39.571024] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.888 [2024-10-17 17:49:39.574537] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.888 [2024-10-17 17:49:39.583643] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.888 [2024-10-17 17:49:39.584326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.888 [2024-10-17 17:49:39.584388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.888 [2024-10-17 17:49:39.584401] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.888 [2024-10-17 17:49:39.584662] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.888 [2024-10-17 17:49:39.584897] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.888 [2024-10-17 17:49:39.584908] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.888 [2024-10-17 17:49:39.584916] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.888 [2024-10-17 17:49:39.588433] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.888 [2024-10-17 17:49:39.597526] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.888 [2024-10-17 17:49:39.598080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.888 [2024-10-17 17:49:39.598106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.888 [2024-10-17 17:49:39.598114] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.888 [2024-10-17 17:49:39.598333] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.888 [2024-10-17 17:49:39.598551] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.888 [2024-10-17 17:49:39.598561] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.888 [2024-10-17 17:49:39.598569] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.888 [2024-10-17 17:49:39.602088] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.888 [2024-10-17 17:49:39.611359] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.888 [2024-10-17 17:49:39.612123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.888 [2024-10-17 17:49:39.612186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.612200] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.612453] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.612677] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.612686] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.612709] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.616247] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.625120] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.625760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.625789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.625799] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.626020] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.626237] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.626247] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.626266] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.629785] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.639069] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.639675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.639751] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.639763] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.640017] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.640239] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.640250] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.640259] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.643778] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.652858] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.653514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.653577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.653589] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.653855] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.654080] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.654093] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.654102] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.657618] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.666720] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.667303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.667330] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.667339] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.667559] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.667790] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.667801] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.667809] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.671327] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.680627] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.681221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.681247] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.681255] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.681474] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.681739] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.681752] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.681760] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.685283] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.694590] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.695132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.695155] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.695164] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.695382] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.695599] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.695611] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.695619] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.699342] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.708462] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.709064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.709091] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.709101] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.709319] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.709539] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.709547] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.709556] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.713077] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.722389] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.722942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.722967] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.722979] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.723198] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.723424] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.723434] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.723443] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.726959] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.736251] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.736807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.736830] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.736839] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.737057] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.737275] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.737285] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.737293] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.740803] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.750091] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.750617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.889 [2024-10-17 17:49:39.750639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.889 [2024-10-17 17:49:39.750648] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.889 [2024-10-17 17:49:39.750875] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.889 [2024-10-17 17:49:39.751093] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.889 [2024-10-17 17:49:39.751110] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.889 [2024-10-17 17:49:39.751120] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.889 [2024-10-17 17:49:39.754625] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.889 [2024-10-17 17:49:39.763930] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.889 [2024-10-17 17:49:39.764442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.890 [2024-10-17 17:49:39.764464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.890 [2024-10-17 17:49:39.764472] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.890 [2024-10-17 17:49:39.764690] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.890 [2024-10-17 17:49:39.764917] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.890 [2024-10-17 17:49:39.764926] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.890 [2024-10-17 17:49:39.764933] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.890 [2024-10-17 17:49:39.768454] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.890 [2024-10-17 17:49:39.777765] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.890 [2024-10-17 17:49:39.778419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.890 [2024-10-17 17:49:39.778481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.890 [2024-10-17 17:49:39.778494] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.890 [2024-10-17 17:49:39.778761] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.890 [2024-10-17 17:49:39.778986] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.890 [2024-10-17 17:49:39.778995] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.890 [2024-10-17 17:49:39.779004] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.890 [2024-10-17 17:49:39.782552] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.890 [2024-10-17 17:49:39.791660] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:31.890 [2024-10-17 17:49:39.792255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:31.890 [2024-10-17 17:49:39.792285] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:31.890 [2024-10-17 17:49:39.792295] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:31.890 [2024-10-17 17:49:39.792515] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:31.890 [2024-10-17 17:49:39.792743] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:31.890 [2024-10-17 17:49:39.792754] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:31.890 [2024-10-17 17:49:39.792762] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:31.890 [2024-10-17 17:49:39.796282] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:31.890 [2024-10-17 17:49:39.805605] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.152 [2024-10-17 17:49:39.806288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.152 [2024-10-17 17:49:39.806352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.152 [2024-10-17 17:49:39.806365] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.152 [2024-10-17 17:49:39.806618] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.152 [2024-10-17 17:49:39.806857] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.152 [2024-10-17 17:49:39.806869] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.152 [2024-10-17 17:49:39.806877] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.152 [2024-10-17 17:49:39.810394] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.152 [2024-10-17 17:49:39.819522] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.152 [2024-10-17 17:49:39.820096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.152 [2024-10-17 17:49:39.820125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.152 [2024-10-17 17:49:39.820142] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.152 [2024-10-17 17:49:39.820362] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.152 [2024-10-17 17:49:39.820580] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.152 [2024-10-17 17:49:39.820589] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.152 [2024-10-17 17:49:39.820597] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.152 [2024-10-17 17:49:39.824123] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.152 [2024-10-17 17:49:39.833422] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.152 [2024-10-17 17:49:39.833986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.152 [2024-10-17 17:49:39.834010] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.152 [2024-10-17 17:49:39.834019] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.152 [2024-10-17 17:49:39.834237] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.152 [2024-10-17 17:49:39.834454] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.152 [2024-10-17 17:49:39.834464] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.152 [2024-10-17 17:49:39.834472] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.152 [2024-10-17 17:49:39.837988] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.152 [2024-10-17 17:49:39.847285] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.152 [2024-10-17 17:49:39.847838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.152 [2024-10-17 17:49:39.847859] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.152 [2024-10-17 17:49:39.847865] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.152 [2024-10-17 17:49:39.848017] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.152 [2024-10-17 17:49:39.848168] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.152 [2024-10-17 17:49:39.848174] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.152 [2024-10-17 17:49:39.848180] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.152 [2024-10-17 17:49:39.850587] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.152 [2024-10-17 17:49:39.859936] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.152 [2024-10-17 17:49:39.860427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.152 [2024-10-17 17:49:39.860444] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.152 [2024-10-17 17:49:39.860451] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.152 [2024-10-17 17:49:39.860602] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.152 [2024-10-17 17:49:39.860760] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.152 [2024-10-17 17:49:39.860779] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.860785] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.863197] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.872536] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.872990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.873006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.873012] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.873162] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.873311] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.873318] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.873323] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.875733] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.885226] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.885655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.885672] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.885677] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.885835] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.885985] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.885991] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.885996] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.888397] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.897914] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.898346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.898366] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.898372] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.898522] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.898671] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.898677] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.898683] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.901091] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.910557] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.911154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.911192] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.911200] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.911370] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.911522] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.911529] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.911535] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.913952] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.923136] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.923699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.923736] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.923745] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.923916] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.924068] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.924075] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.924080] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.926486] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.935793] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.936349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.936385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.936393] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.936561] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.936722] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.936730] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.936735] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.939135] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.948440] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.949006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.949039] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.949048] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.949218] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.949370] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.949378] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.949384] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.951787] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.961100] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.961654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.961687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.961703] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.961871] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.962023] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.962029] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.962035] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.964434] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.973838] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.974298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.974312] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.974318] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.974468] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.974616] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.974622] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.974627] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.977027] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.986477] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.987050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.987081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.987090] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.987254] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.153 [2024-10-17 17:49:39.987405] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.153 [2024-10-17 17:49:39.987412] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.153 [2024-10-17 17:49:39.987421] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.153 [2024-10-17 17:49:39.989829] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.153 [2024-10-17 17:49:39.999136] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.153 [2024-10-17 17:49:39.999661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.153 [2024-10-17 17:49:39.999698] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.153 [2024-10-17 17:49:39.999707] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.153 [2024-10-17 17:49:39.999871] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.154 [2024-10-17 17:49:40.000022] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.154 [2024-10-17 17:49:40.000029] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.154 [2024-10-17 17:49:40.000034] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.154 [2024-10-17 17:49:40.002895] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.154 [2024-10-17 17:49:40.011836] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.154 [2024-10-17 17:49:40.012396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.154 [2024-10-17 17:49:40.012426] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.154 [2024-10-17 17:49:40.012435] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.154 [2024-10-17 17:49:40.012599] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.154 [2024-10-17 17:49:40.012758] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.154 [2024-10-17 17:49:40.012765] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.154 [2024-10-17 17:49:40.012771] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.154 [2024-10-17 17:49:40.015167] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.154 [2024-10-17 17:49:40.024479] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.154 [2024-10-17 17:49:40.025057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.154 [2024-10-17 17:49:40.025087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.154 [2024-10-17 17:49:40.025097] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.154 [2024-10-17 17:49:40.025261] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.154 [2024-10-17 17:49:40.025413] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.154 [2024-10-17 17:49:40.025419] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.154 [2024-10-17 17:49:40.025425] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.154 [2024-10-17 17:49:40.027829] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.154 [2024-10-17 17:49:40.037113] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.154 [2024-10-17 17:49:40.037649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.154 [2024-10-17 17:49:40.037679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.154 [2024-10-17 17:49:40.037688] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.154 [2024-10-17 17:49:40.037859] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.154 [2024-10-17 17:49:40.038011] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.154 [2024-10-17 17:49:40.038017] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.154 [2024-10-17 17:49:40.038023] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.154 [2024-10-17 17:49:40.040419] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.154 [2024-10-17 17:49:40.049726] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.154 [2024-10-17 17:49:40.050285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.154 [2024-10-17 17:49:40.050315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.154 [2024-10-17 17:49:40.050324] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.154 [2024-10-17 17:49:40.050489] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.154 [2024-10-17 17:49:40.050640] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.154 [2024-10-17 17:49:40.050646] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.154 [2024-10-17 17:49:40.050652] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.154 [2024-10-17 17:49:40.053057] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.154 [2024-10-17 17:49:40.062363] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.154 [2024-10-17 17:49:40.062999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.154 [2024-10-17 17:49:40.063029] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.154 [2024-10-17 17:49:40.063038] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.154 [2024-10-17 17:49:40.063202] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.154 [2024-10-17 17:49:40.063353] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.154 [2024-10-17 17:49:40.063361] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.154 [2024-10-17 17:49:40.063366] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.154 [2024-10-17 17:49:40.065764] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.415 [2024-10-17 17:49:40.074961] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.415 [2024-10-17 17:49:40.075423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.415 [2024-10-17 17:49:40.075438] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.415 [2024-10-17 17:49:40.075444] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.415 [2024-10-17 17:49:40.075593] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.415 [2024-10-17 17:49:40.075752] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.415 [2024-10-17 17:49:40.075759] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.415 [2024-10-17 17:49:40.075764] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.415 [2024-10-17 17:49:40.078156] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.415 [2024-10-17 17:49:40.087609] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.415 [2024-10-17 17:49:40.088155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.415 [2024-10-17 17:49:40.088186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.415 [2024-10-17 17:49:40.088195] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.415 [2024-10-17 17:49:40.088358] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.415 [2024-10-17 17:49:40.088510] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.415 [2024-10-17 17:49:40.088516] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.415 [2024-10-17 17:49:40.088521] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.415 [2024-10-17 17:49:40.090924] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.415 [2024-10-17 17:49:40.100262] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.415 [2024-10-17 17:49:40.100826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.415 [2024-10-17 17:49:40.100857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.415 [2024-10-17 17:49:40.100865] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.415 [2024-10-17 17:49:40.101032] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.415 [2024-10-17 17:49:40.101184] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.101191] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.101197] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.103598] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.112899] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.113483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.113513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.113522] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.113686] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.113845] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.113851] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.113857] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.116258] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.125593] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.126074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.126089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.126095] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.126244] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.126392] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.126398] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.126403] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.128798] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.138240] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.138714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.138745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.138754] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.138921] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.139072] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.139078] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.139084] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.141485] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.150923] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.151471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.151502] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.151510] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.151674] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.151833] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.151840] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.151846] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.154244] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.163541] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.163988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.164003] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.164012] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.164161] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.164309] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.164315] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.164320] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.166715] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.176148] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.176707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.176737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.176746] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.176913] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.177064] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.177071] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.177076] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.179479] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.188790] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.189260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.189274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.189280] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.189429] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.189578] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.189583] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.189588] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.191983] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.201424] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.202000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.202031] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.202040] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.202206] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.202358] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.202367] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.202373] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.204781] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.214005] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.214535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.214565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.214574] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.214746] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.214898] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.214905] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.214910] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.217315] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.226623] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.227093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.227108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.227114] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.416 [2024-10-17 17:49:40.227263] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.416 [2024-10-17 17:49:40.227411] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.416 [2024-10-17 17:49:40.227416] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.416 [2024-10-17 17:49:40.227422] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.416 [2024-10-17 17:49:40.229817] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.416 [2024-10-17 17:49:40.239256] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.416 [2024-10-17 17:49:40.239760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.416 [2024-10-17 17:49:40.239791] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.416 [2024-10-17 17:49:40.239799] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.239966] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.240118] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.240124] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.240130] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.242531] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.417 [2024-10-17 17:49:40.251837] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.417 [2024-10-17 17:49:40.252382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.417 [2024-10-17 17:49:40.252412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.417 [2024-10-17 17:49:40.252421] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.252585] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.252743] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.252750] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.252756] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.255153] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.417 [2024-10-17 17:49:40.264452] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.417 [2024-10-17 17:49:40.265011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.417 [2024-10-17 17:49:40.265042] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.417 [2024-10-17 17:49:40.265051] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.265215] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.265366] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.265372] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.265377] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.267780] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.417 [2024-10-17 17:49:40.277086] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.417 [2024-10-17 17:49:40.277547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.417 [2024-10-17 17:49:40.277561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.417 [2024-10-17 17:49:40.277567] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.277720] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.277869] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.277875] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.277880] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.280271] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.417 [2024-10-17 17:49:40.289714] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.417 [2024-10-17 17:49:40.290164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.417 [2024-10-17 17:49:40.290176] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.417 [2024-10-17 17:49:40.290181] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.290333] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.290482] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.290488] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.290493] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.292886] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.417 [2024-10-17 17:49:40.302332] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.417 [2024-10-17 17:49:40.302871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.417 [2024-10-17 17:49:40.302901] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.417 [2024-10-17 17:49:40.302910] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.303074] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.303225] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.303232] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.303237] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.305640] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.417 [2024-10-17 17:49:40.314976] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.417 [2024-10-17 17:49:40.315441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.417 [2024-10-17 17:49:40.315455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.417 [2024-10-17 17:49:40.315461] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.315610] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.315763] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.315769] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.315775] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.318173] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.417 [2024-10-17 17:49:40.327612] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.417 [2024-10-17 17:49:40.328155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.417 [2024-10-17 17:49:40.328185] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.417 [2024-10-17 17:49:40.328194] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.417 [2024-10-17 17:49:40.328358] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.417 [2024-10-17 17:49:40.328509] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.417 [2024-10-17 17:49:40.328516] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.417 [2024-10-17 17:49:40.328525] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.417 [2024-10-17 17:49:40.330930] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.340236] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.340807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.340837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.340845] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.341010] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.341161] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.341167] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.341173] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.343575] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.352877] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.353398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.353428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.353437] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.353601] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.353760] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.353767] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.353773] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.356168] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.365466] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.366027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.366057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.366065] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.366229] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.366381] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.366387] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.366392] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.368794] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.378094] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.378442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.378456] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.378462] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.378611] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.378765] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.378772] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.378777] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.381178] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.390757] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.391276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.391306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.391315] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.391479] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.391630] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.391637] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.391642] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.394045] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.403365] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.403899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.403929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.403937] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.404102] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.404253] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.404259] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.404265] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.406669] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 6546.00 IOPS, 25.57 MiB/s [2024-10-17T15:49:40.598Z] [2024-10-17 17:49:40.415996] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.416318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.416333] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.416339] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.416493] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.416642] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.416647] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.416652] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.419062] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.428650] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.429180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.429210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.429219] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.679 [2024-10-17 17:49:40.429383] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.679 [2024-10-17 17:49:40.429534] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.679 [2024-10-17 17:49:40.429540] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.679 [2024-10-17 17:49:40.429546] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.679 [2024-10-17 17:49:40.431945] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.679 [2024-10-17 17:49:40.441252] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.679 [2024-10-17 17:49:40.441812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.679 [2024-10-17 17:49:40.441847] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.679 [2024-10-17 17:49:40.441856] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.442023] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.442174] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.442180] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.442185] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.444588] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.453895] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.454464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.454495] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.454503] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.454667] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.454827] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.454834] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.454844] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.457242] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.466589] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.467148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.467178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.467187] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.467351] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.467502] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.467508] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.467514] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.469921] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.479224] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.479731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.479752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.479758] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.479912] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.480062] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.480067] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.480073] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.482478] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.491920] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.492437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.492468] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.492476] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.492640] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.492799] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.492806] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.492812] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.495210] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.504514] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.505082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.505120] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.505128] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.505292] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.505443] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.505449] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.505454] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.507858] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.517182] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.517711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.517741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.517750] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.517914] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.518072] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.518079] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.518085] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.520490] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.529790] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.530322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.530352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.530360] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.530525] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.530676] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.530682] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.530687] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.533095] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.542405] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.542908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.542938] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.542947] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.543111] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.543266] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.543272] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.543277] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.545681] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.554985] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.555510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.555540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.555548] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.555720] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.555872] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.555878] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.555884] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.558279] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.567578] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.568134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.680 [2024-10-17 17:49:40.568165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.680 [2024-10-17 17:49:40.568173] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.680 [2024-10-17 17:49:40.568337] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.680 [2024-10-17 17:49:40.568488] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.680 [2024-10-17 17:49:40.568495] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.680 [2024-10-17 17:49:40.568500] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.680 [2024-10-17 17:49:40.570903] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.680 [2024-10-17 17:49:40.580202] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.680 [2024-10-17 17:49:40.580661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.681 [2024-10-17 17:49:40.580675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.681 [2024-10-17 17:49:40.580681] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.681 [2024-10-17 17:49:40.580840] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.681 [2024-10-17 17:49:40.580990] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.681 [2024-10-17 17:49:40.580996] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.681 [2024-10-17 17:49:40.581001] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.681 [2024-10-17 17:49:40.583392] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.681 [2024-10-17 17:49:40.592834] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.681 [2024-10-17 17:49:40.593283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.681 [2024-10-17 17:49:40.593296] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.681 [2024-10-17 17:49:40.593301] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.681 [2024-10-17 17:49:40.593449] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.681 [2024-10-17 17:49:40.593597] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.681 [2024-10-17 17:49:40.593603] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.681 [2024-10-17 17:49:40.593608] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.681 [2024-10-17 17:49:40.596027] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.942 [2024-10-17 17:49:40.605477] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.942 [2024-10-17 17:49:40.605940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.942 [2024-10-17 17:49:40.605970] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.942 [2024-10-17 17:49:40.605979] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.942 [2024-10-17 17:49:40.606143] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.942 [2024-10-17 17:49:40.606294] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.942 [2024-10-17 17:49:40.606301] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.942 [2024-10-17 17:49:40.606306] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.942 [2024-10-17 17:49:40.608709] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.942 [2024-10-17 17:49:40.618151] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.942 [2024-10-17 17:49:40.618683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.942 [2024-10-17 17:49:40.618719] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.942 [2024-10-17 17:49:40.618728] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.942 [2024-10-17 17:49:40.618892] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.942 [2024-10-17 17:49:40.619044] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.942 [2024-10-17 17:49:40.619050] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.942 [2024-10-17 17:49:40.619055] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.942 [2024-10-17 17:49:40.621452] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.630758] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.631301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.631332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.631343] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.631507] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.631659] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.631665] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.631670] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.634073] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.643383] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.643837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.643852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.643858] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.644007] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.644155] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.644160] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.644165] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.646560] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.656008] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.656548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.656578] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.656587] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.656759] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.656913] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.656920] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.656925] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.659322] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.668631] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.669208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.669238] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.669247] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.669411] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.669562] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.669572] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.669577] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.671985] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.681312] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.681901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.681932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.681940] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.682104] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.682255] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.682262] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.682267] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.684662] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.693963] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.694454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.694468] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.694474] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.694622] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.694775] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.694782] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.694787] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.697177] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.706619] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.707071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.707084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.707089] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.707238] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.707387] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.707392] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.707398] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.709792] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.719238] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.719696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.719709] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.719714] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.719863] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.720011] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.720017] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.720023] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.722594] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.731914] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.732448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.732478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.732487] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.732651] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.732812] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.732819] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.732825] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.735220] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.744530] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.745079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.745110] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.745118] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.745282] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.745433] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.943 [2024-10-17 17:49:40.745440] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.943 [2024-10-17 17:49:40.745445] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.943 [2024-10-17 17:49:40.747850] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.943 [2024-10-17 17:49:40.757160] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.943 [2024-10-17 17:49:40.757625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.943 [2024-10-17 17:49:40.757639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.943 [2024-10-17 17:49:40.757644] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.943 [2024-10-17 17:49:40.757803] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.943 [2024-10-17 17:49:40.757951] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.757957] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.757962] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.760353] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.769798] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.770255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.770267] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.770272] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.770420] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.770568] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.770574] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.770579] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.772980] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.782433] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.782910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.782940] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.782949] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.783116] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.783267] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.783273] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.783278] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.785679] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.795027] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.795576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.795606] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.795615] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.795787] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.795939] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.795946] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.795955] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.798360] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.807681] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.808050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.808066] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.808071] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.808220] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.808368] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.808373] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.808379] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.810776] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.820376] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.820823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.820836] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.820842] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.820991] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.821139] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.821145] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.821150] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.823543] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.832993] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.833419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.833449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.833458] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.833622] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.833782] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.833790] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.833795] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.836192] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.845639] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.846186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.846220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.846228] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.846393] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.846544] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.846550] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.846555] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:32.944 [2024-10-17 17:49:40.848964] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:32.944 [2024-10-17 17:49:40.858278] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:32.944 [2024-10-17 17:49:40.858817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:32.944 [2024-10-17 17:49:40.858848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:32.944 [2024-10-17 17:49:40.858856] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:32.944 [2024-10-17 17:49:40.859020] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:32.944 [2024-10-17 17:49:40.859172] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:32.944 [2024-10-17 17:49:40.859178] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:32.944 [2024-10-17 17:49:40.859183] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.206 [2024-10-17 17:49:40.861587] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.206 [2024-10-17 17:49:40.870892] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.206 [2024-10-17 17:49:40.871415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.206 [2024-10-17 17:49:40.871446] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.206 [2024-10-17 17:49:40.871454] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.206 [2024-10-17 17:49:40.871619] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.206 [2024-10-17 17:49:40.871779] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.206 [2024-10-17 17:49:40.871786] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.206 [2024-10-17 17:49:40.871792] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.206 [2024-10-17 17:49:40.874185] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.206 [2024-10-17 17:49:40.883496] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.206 [2024-10-17 17:49:40.883953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.206 [2024-10-17 17:49:40.883967] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.206 [2024-10-17 17:49:40.883973] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.206 [2024-10-17 17:49:40.884122] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.206 [2024-10-17 17:49:40.884275] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.206 [2024-10-17 17:49:40.884281] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.206 [2024-10-17 17:49:40.884286] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.206 [2024-10-17 17:49:40.886681] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.206 [2024-10-17 17:49:40.896133] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.206 [2024-10-17 17:49:40.896578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.206 [2024-10-17 17:49:40.896590] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.206 [2024-10-17 17:49:40.896596] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.206 [2024-10-17 17:49:40.896749] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.206 [2024-10-17 17:49:40.896898] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.206 [2024-10-17 17:49:40.896904] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.206 [2024-10-17 17:49:40.896910] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.206 [2024-10-17 17:49:40.899309] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.206 [2024-10-17 17:49:40.908756] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.206 [2024-10-17 17:49:40.909195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.206 [2024-10-17 17:49:40.909226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.206 [2024-10-17 17:49:40.909234] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.206 [2024-10-17 17:49:40.909399] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.206 [2024-10-17 17:49:40.909550] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.206 [2024-10-17 17:49:40.909556] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.206 [2024-10-17 17:49:40.909561] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.206 [2024-10-17 17:49:40.911963] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.206 [2024-10-17 17:49:40.921419] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.206 [2024-10-17 17:49:40.921994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.206 [2024-10-17 17:49:40.922024] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.206 [2024-10-17 17:49:40.922033] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.206 [2024-10-17 17:49:40.922197] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.206 [2024-10-17 17:49:40.922349] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.206 [2024-10-17 17:49:40.922355] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.206 [2024-10-17 17:49:40.922360] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.206 [2024-10-17 17:49:40.924761] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.206 [2024-10-17 17:49:40.934102] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.206 [2024-10-17 17:49:40.934518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.206 [2024-10-17 17:49:40.934533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.206 [2024-10-17 17:49:40.934539] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.206 [2024-10-17 17:49:40.934687] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.206 [2024-10-17 17:49:40.934841] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.206 [2024-10-17 17:49:40.934847] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.206 [2024-10-17 17:49:40.934852] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.206 [2024-10-17 17:49:40.937242] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.206 [2024-10-17 17:49:40.946685] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.206 [2024-10-17 17:49:40.947139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.206 [2024-10-17 17:49:40.947151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.206 [2024-10-17 17:49:40.947157] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.206 [2024-10-17 17:49:40.947305] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.206 [2024-10-17 17:49:40.947454] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.206 [2024-10-17 17:49:40.947460] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.206 [2024-10-17 17:49:40.947465] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:40.949860] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:40.959308] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:40.959792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:40.959823] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:40.959832] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:40.959998] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:40.960150] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:40.960156] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:40.960161] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:40.962564] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:40.971881] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:40.972423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:40.972453] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:40.972465] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:40.972630] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:40.972787] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:40.972794] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:40.972799] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:40.975195] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:40.984511] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:40.985037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:40.985068] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:40.985076] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:40.985241] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:40.985392] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:40.985398] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:40.985403] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:40.987803] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:40.997109] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:40.997565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:40.997579] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:40.997585] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:40.997739] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:40.997891] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:40.997899] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:40.997904] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:41.000297] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:41.009687] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:41.010159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:41.010173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:41.010178] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:41.010326] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:41.010475] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:41.010484] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:41.010490] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:41.012885] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:41.022344] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:41.022820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:41.022851] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:41.022859] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:41.023026] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:41.023177] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:41.023183] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:41.023188] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:41.025592] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:41.034912] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:41.035458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:41.035489] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:41.035497] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:41.035661] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:41.035819] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:41.035827] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:41.035833] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:41.038230] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:41.047540] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:41.048025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:41.048040] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:41.048045] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:41.048194] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:41.048343] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:41.048349] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:41.048354] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:41.050749] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:41.060196] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:41.060711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:41.060742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:41.060750] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:41.060914] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:41.061065] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:41.061071] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:41.061076] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:41.063476] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:41.072789] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:41.073252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.207 [2024-10-17 17:49:41.073266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.207 [2024-10-17 17:49:41.073272] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.207 [2024-10-17 17:49:41.073420] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.207 [2024-10-17 17:49:41.073568] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.207 [2024-10-17 17:49:41.073574] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.207 [2024-10-17 17:49:41.073579] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.207 [2024-10-17 17:49:41.075977] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.207 [2024-10-17 17:49:41.085429] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.207 [2024-10-17 17:49:41.085917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.208 [2024-10-17 17:49:41.085930] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.208 [2024-10-17 17:49:41.085935] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.208 [2024-10-17 17:49:41.086084] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.208 [2024-10-17 17:49:41.086232] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.208 [2024-10-17 17:49:41.086237] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.208 [2024-10-17 17:49:41.086242] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.208 [2024-10-17 17:49:41.088633] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.208 [2024-10-17 17:49:41.098084] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.208 [2024-10-17 17:49:41.098543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.208 [2024-10-17 17:49:41.098554] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.208 [2024-10-17 17:49:41.098560] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.208 [2024-10-17 17:49:41.098715] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.208 [2024-10-17 17:49:41.098864] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.208 [2024-10-17 17:49:41.098870] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.208 [2024-10-17 17:49:41.098875] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.208 [2024-10-17 17:49:41.101267] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.208 [2024-10-17 17:49:41.110743] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.208 [2024-10-17 17:49:41.111223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.208 [2024-10-17 17:49:41.111235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.208 [2024-10-17 17:49:41.111240] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.208 [2024-10-17 17:49:41.111388] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.208 [2024-10-17 17:49:41.111536] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.208 [2024-10-17 17:49:41.111542] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.208 [2024-10-17 17:49:41.111547] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.208 [2024-10-17 17:49:41.113940] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.208 [2024-10-17 17:49:41.123392] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.208 [2024-10-17 17:49:41.123735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.208 [2024-10-17 17:49:41.123748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.208 [2024-10-17 17:49:41.123753] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.123901] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.124051] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.124057] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.124062] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.470 [2024-10-17 17:49:41.126453] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.470 [2024-10-17 17:49:41.136036] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.470 [2024-10-17 17:49:41.136537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.470 [2024-10-17 17:49:41.136549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.470 [2024-10-17 17:49:41.136554] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.136706] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.136855] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.136860] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.136872] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.470 [2024-10-17 17:49:41.139286] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.470 [2024-10-17 17:49:41.148601] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.470 [2024-10-17 17:49:41.149197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.470 [2024-10-17 17:49:41.149227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.470 [2024-10-17 17:49:41.149236] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.149400] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.149552] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.149558] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.149563] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.470 [2024-10-17 17:49:41.151962] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.470 [2024-10-17 17:49:41.161267] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.470 [2024-10-17 17:49:41.161726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.470 [2024-10-17 17:49:41.161741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.470 [2024-10-17 17:49:41.161747] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.161895] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.162044] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.162049] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.162054] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.470 [2024-10-17 17:49:41.164447] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.470 [2024-10-17 17:49:41.173894] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.470 [2024-10-17 17:49:41.174343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.470 [2024-10-17 17:49:41.174355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.470 [2024-10-17 17:49:41.174360] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.174508] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.174656] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.174662] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.174667] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.470 [2024-10-17 17:49:41.177061] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.470 [2024-10-17 17:49:41.186504] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.470 [2024-10-17 17:49:41.186959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.470 [2024-10-17 17:49:41.186992] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.470 [2024-10-17 17:49:41.187000] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.187166] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.187317] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.187323] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.187328] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.470 [2024-10-17 17:49:41.189729] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.470 [2024-10-17 17:49:41.199176] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.470 [2024-10-17 17:49:41.199607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.470 [2024-10-17 17:49:41.199621] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.470 [2024-10-17 17:49:41.199627] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.199781] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.199930] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.199935] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.199940] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.470 [2024-10-17 17:49:41.202331] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.470 [2024-10-17 17:49:41.211774] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.470 [2024-10-17 17:49:41.212219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.470 [2024-10-17 17:49:41.212231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.470 [2024-10-17 17:49:41.212236] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.470 [2024-10-17 17:49:41.212385] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.470 [2024-10-17 17:49:41.212533] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.470 [2024-10-17 17:49:41.212538] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.470 [2024-10-17 17:49:41.212544] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.214939] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.224383] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.224736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.224749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.224754] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.224903] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.225055] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.225069] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.225074] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.227466] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.237045] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.237590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.237620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.237629] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.237799] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.237951] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.237958] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.237963] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.240360] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.249662] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.250217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.250247] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.250256] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.250420] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.250571] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.250577] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.250582] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.252984] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.262292] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.262829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.262859] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.262868] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.263035] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.263186] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.263193] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.263198] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.265601] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.274913] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.275370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.275385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.275390] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.275539] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.275688] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.275698] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.275703] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.278096] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.287546] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.288019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.288031] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.288037] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.288185] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.288333] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.288339] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.288344] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.290737] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.300204] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.300661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.300673] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.300678] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.300831] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.300980] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.300985] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.300991] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.303381] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.312829] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.313307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.313338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.471 [2024-10-17 17:49:41.313351] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.471 [2024-10-17 17:49:41.313515] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.471 [2024-10-17 17:49:41.313667] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.471 [2024-10-17 17:49:41.313674] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.471 [2024-10-17 17:49:41.313679] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.471 [2024-10-17 17:49:41.316085] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.471 [2024-10-17 17:49:41.325398] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.471 [2024-10-17 17:49:41.325916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.471 [2024-10-17 17:49:41.325947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.472 [2024-10-17 17:49:41.325955] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.472 [2024-10-17 17:49:41.326119] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.472 [2024-10-17 17:49:41.326271] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.472 [2024-10-17 17:49:41.326277] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.472 [2024-10-17 17:49:41.326283] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.472 [2024-10-17 17:49:41.328681] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.472 [2024-10-17 17:49:41.337992] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.472 [2024-10-17 17:49:41.338540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.472 [2024-10-17 17:49:41.338570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.472 [2024-10-17 17:49:41.338579] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.472 [2024-10-17 17:49:41.338749] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.472 [2024-10-17 17:49:41.338901] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.472 [2024-10-17 17:49:41.338907] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.472 [2024-10-17 17:49:41.338912] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.472 [2024-10-17 17:49:41.341304] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.472 [2024-10-17 17:49:41.350638] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.472 [2024-10-17 17:49:41.351190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.472 [2024-10-17 17:49:41.351220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.472 [2024-10-17 17:49:41.351228] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.472 [2024-10-17 17:49:41.351392] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.472 [2024-10-17 17:49:41.351544] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.472 [2024-10-17 17:49:41.351554] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.472 [2024-10-17 17:49:41.351560] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.472 [2024-10-17 17:49:41.353962] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.472 [2024-10-17 17:49:41.363266] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.472 [2024-10-17 17:49:41.363735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.472 [2024-10-17 17:49:41.363749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.472 [2024-10-17 17:49:41.363755] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.472 [2024-10-17 17:49:41.363904] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.472 [2024-10-17 17:49:41.364052] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.472 [2024-10-17 17:49:41.364058] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.472 [2024-10-17 17:49:41.364063] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.472 [2024-10-17 17:49:41.366456] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.472 [2024-10-17 17:49:41.375899] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.472 [2024-10-17 17:49:41.376350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.472 [2024-10-17 17:49:41.376362] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.472 [2024-10-17 17:49:41.376367] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.472 [2024-10-17 17:49:41.376515] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.472 [2024-10-17 17:49:41.376663] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.472 [2024-10-17 17:49:41.376668] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.472 [2024-10-17 17:49:41.376673] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.472 [2024-10-17 17:49:41.379066] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.734 [2024-10-17 17:49:41.388516] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.734 [2024-10-17 17:49:41.388849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.734 [2024-10-17 17:49:41.388862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.734 [2024-10-17 17:49:41.388867] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.734 [2024-10-17 17:49:41.389016] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.734 [2024-10-17 17:49:41.389164] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.734 [2024-10-17 17:49:41.389171] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.734 [2024-10-17 17:49:41.389176] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.734 [2024-10-17 17:49:41.391571] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.734 [2024-10-17 17:49:41.401162] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.734 [2024-10-17 17:49:41.401712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.734 [2024-10-17 17:49:41.401743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.734 [2024-10-17 17:49:41.401751] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.734 [2024-10-17 17:49:41.401918] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.734 [2024-10-17 17:49:41.402069] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.734 [2024-10-17 17:49:41.402076] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.734 [2024-10-17 17:49:41.402081] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.734 [2024-10-17 17:49:41.404482] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.734 5236.80 IOPS, 20.46 MiB/s [2024-10-17T15:49:41.653Z] [2024-10-17 17:49:41.414925] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.734 [2024-10-17 17:49:41.415384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.734 [2024-10-17 17:49:41.415399] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.734 [2024-10-17 17:49:41.415404] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.734 [2024-10-17 17:49:41.415553] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.734 [2024-10-17 17:49:41.415707] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.734 [2024-10-17 17:49:41.415714] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.734 [2024-10-17 17:49:41.415719] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.734 [2024-10-17 17:49:41.418111] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.734 [2024-10-17 17:49:41.427561] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.734 [2024-10-17 17:49:41.428150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.734 [2024-10-17 17:49:41.428180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.734 [2024-10-17 17:49:41.428188] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.734 [2024-10-17 17:49:41.428352] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.734 [2024-10-17 17:49:41.428504] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.734 [2024-10-17 17:49:41.428510] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.734 [2024-10-17 17:49:41.428516] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.734 [2024-10-17 17:49:41.430920] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.734 [2024-10-17 17:49:41.440232] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.734 [2024-10-17 17:49:41.440787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.734 [2024-10-17 17:49:41.440818] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.734 [2024-10-17 17:49:41.440827] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.734 [2024-10-17 17:49:41.440997] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.734 [2024-10-17 17:49:41.441149] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.734 [2024-10-17 17:49:41.441155] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.441161] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.443561] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.452877] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.453339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.453353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.453359] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.453507] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.453656] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.453662] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.453667] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.456064] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.465544] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.466095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.466126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.466135] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.466299] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.466451] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.466457] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.466463] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.468867] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.478178] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.478640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.478655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.478660] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.478818] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.478967] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.478973] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.478982] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.481377] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.490834] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.491304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.491316] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.491322] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.491470] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.491618] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.491624] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.491629] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.494028] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.503479] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.504081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.504111] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.504120] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.504285] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.504436] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.504442] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.504447] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.506851] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.516159] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.516726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.516756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.516765] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.516930] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.517081] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.517087] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.517093] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.519496] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.528814] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.529293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.529327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.529336] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.529501] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.529652] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.529659] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.529664] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.532063] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.541502] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.541950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.541965] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.541970] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.542119] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.542267] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.542273] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.542278] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.544668] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.554113] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.554575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.554587] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.735 [2024-10-17 17:49:41.554592] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.735 [2024-10-17 17:49:41.554765] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.735 [2024-10-17 17:49:41.554916] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.735 [2024-10-17 17:49:41.554922] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.735 [2024-10-17 17:49:41.554927] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.735 [2024-10-17 17:49:41.557321] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.735 [2024-10-17 17:49:41.566771] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.735 [2024-10-17 17:49:41.567207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.735 [2024-10-17 17:49:41.567220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.736 [2024-10-17 17:49:41.567225] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.736 [2024-10-17 17:49:41.567374] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.736 [2024-10-17 17:49:41.567526] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.736 [2024-10-17 17:49:41.567531] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.736 [2024-10-17 17:49:41.567536] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.736 [2024-10-17 17:49:41.569931] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.736 [2024-10-17 17:49:41.579364] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.736 [2024-10-17 17:49:41.579702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.736 [2024-10-17 17:49:41.579714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.736 [2024-10-17 17:49:41.579720] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.736 [2024-10-17 17:49:41.579868] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.736 [2024-10-17 17:49:41.580015] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.736 [2024-10-17 17:49:41.580021] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.736 [2024-10-17 17:49:41.580026] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.736 [2024-10-17 17:49:41.582420] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.736 [2024-10-17 17:49:41.592015] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.736 [2024-10-17 17:49:41.592532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.736 [2024-10-17 17:49:41.592562] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.736 [2024-10-17 17:49:41.592571] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.736 [2024-10-17 17:49:41.592742] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.736 [2024-10-17 17:49:41.592894] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.736 [2024-10-17 17:49:41.592901] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.736 [2024-10-17 17:49:41.592906] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.736 [2024-10-17 17:49:41.595301] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.736 [2024-10-17 17:49:41.604613] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.736 [2024-10-17 17:49:41.605095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.736 [2024-10-17 17:49:41.605109] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.736 [2024-10-17 17:49:41.605115] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.736 [2024-10-17 17:49:41.605263] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.736 [2024-10-17 17:49:41.605412] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.736 [2024-10-17 17:49:41.605417] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.736 [2024-10-17 17:49:41.605422] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.736 [2024-10-17 17:49:41.607823] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.736 [2024-10-17 17:49:41.617268] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.736 [2024-10-17 17:49:41.617669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.736 [2024-10-17 17:49:41.617681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.736 [2024-10-17 17:49:41.617687] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.736 [2024-10-17 17:49:41.617840] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.736 [2024-10-17 17:49:41.617989] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.736 [2024-10-17 17:49:41.617995] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.736 [2024-10-17 17:49:41.617999] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.736 [2024-10-17 17:49:41.620394] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.736 [2024-10-17 17:49:41.629847] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.736 [2024-10-17 17:49:41.630385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.736 [2024-10-17 17:49:41.630416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.736 [2024-10-17 17:49:41.630424] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.736 [2024-10-17 17:49:41.630588] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.736 [2024-10-17 17:49:41.630747] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.736 [2024-10-17 17:49:41.630754] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.736 [2024-10-17 17:49:41.630759] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.736 [2024-10-17 17:49:41.633154] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.736 [2024-10-17 17:49:41.642457] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.736 [2024-10-17 17:49:41.643061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.736 [2024-10-17 17:49:41.643092] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.736 [2024-10-17 17:49:41.643101] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.736 [2024-10-17 17:49:41.643265] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.736 [2024-10-17 17:49:41.643416] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.736 [2024-10-17 17:49:41.643422] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.736 [2024-10-17 17:49:41.643427] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.736 [2024-10-17 17:49:41.645831] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.998 [2024-10-17 17:49:41.655136] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.998 [2024-10-17 17:49:41.655687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.998 [2024-10-17 17:49:41.655722] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.998 [2024-10-17 17:49:41.655734] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.998 [2024-10-17 17:49:41.655898] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.998 [2024-10-17 17:49:41.656049] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.998 [2024-10-17 17:49:41.656055] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.998 [2024-10-17 17:49:41.656061] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.998 [2024-10-17 17:49:41.658460] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.998 [2024-10-17 17:49:41.667762] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.998 [2024-10-17 17:49:41.668201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.998 [2024-10-17 17:49:41.668215] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.998 [2024-10-17 17:49:41.668221] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.998 [2024-10-17 17:49:41.668369] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.998 [2024-10-17 17:49:41.668518] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.998 [2024-10-17 17:49:41.668523] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.998 [2024-10-17 17:49:41.668529] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.998 [2024-10-17 17:49:41.670929] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.998 [2024-10-17 17:49:41.680368] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.998 [2024-10-17 17:49:41.680824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.998 [2024-10-17 17:49:41.680837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.998 [2024-10-17 17:49:41.680843] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.998 [2024-10-17 17:49:41.680991] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.998 [2024-10-17 17:49:41.681139] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.998 [2024-10-17 17:49:41.681145] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.998 [2024-10-17 17:49:41.681150] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.998 [2024-10-17 17:49:41.683581] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.998 [2024-10-17 17:49:41.693038] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.998 [2024-10-17 17:49:41.693557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.998 [2024-10-17 17:49:41.693588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.998 [2024-10-17 17:49:41.693596] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.998 [2024-10-17 17:49:41.693768] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.998 [2024-10-17 17:49:41.693921] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.998 [2024-10-17 17:49:41.693930] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.998 [2024-10-17 17:49:41.693936] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.998 [2024-10-17 17:49:41.696332] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.998 [2024-10-17 17:49:41.705633] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.998 [2024-10-17 17:49:41.706183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.998 [2024-10-17 17:49:41.706213] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.998 [2024-10-17 17:49:41.706222] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.998 [2024-10-17 17:49:41.706386] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.998 [2024-10-17 17:49:41.706537] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.998 [2024-10-17 17:49:41.706544] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.998 [2024-10-17 17:49:41.706549] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.998 [2024-10-17 17:49:41.708953] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.998 [2024-10-17 17:49:41.718255] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.998 [2024-10-17 17:49:41.718793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.998 [2024-10-17 17:49:41.718824] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.998 [2024-10-17 17:49:41.718833] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.998 [2024-10-17 17:49:41.718999] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.998 [2024-10-17 17:49:41.719150] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.998 [2024-10-17 17:49:41.719157] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.998 [2024-10-17 17:49:41.719163] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.998 [2024-10-17 17:49:41.721567] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.998 [2024-10-17 17:49:41.730881] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.998 [2024-10-17 17:49:41.731427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.731458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.731466] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.731630] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.731787] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.731794] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.731799] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.734193] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.743496] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.743954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.743969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.743975] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.744124] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.744273] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.744278] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.744284] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.746673] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.756115] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.756653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.756683] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.756699] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.756866] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.757017] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.757023] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.757029] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.759422] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.768747] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.769260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.769290] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.769299] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.769463] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.769614] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.769620] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.769625] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.772027] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.781326] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.781780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.781810] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.781819] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.781992] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.782143] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.782149] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.782154] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.784555] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.794003] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.794532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.794562] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.794570] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.794742] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.794894] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.794901] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.794906] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.797301] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.806609] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.807127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.807157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.807166] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.807330] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.807481] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.807488] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.807493] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.809892] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.819195] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.819747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.819777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.819786] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.819950] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.820101] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.820108] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.820117] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.822527] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.831831] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.832371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.832401] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.832410] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.832574] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.832733] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.832740] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.832746] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.835141] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.844441] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.844917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.844932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.844938] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.845086] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.845235] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.845240] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.845245] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.847637] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:33.999 [2024-10-17 17:49:41.857070] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:33.999 [2024-10-17 17:49:41.857501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:33.999 [2024-10-17 17:49:41.857513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:33.999 [2024-10-17 17:49:41.857518] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:33.999 [2024-10-17 17:49:41.857666] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:33.999 [2024-10-17 17:49:41.857820] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:33.999 [2024-10-17 17:49:41.857826] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:33.999 [2024-10-17 17:49:41.857831] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:33.999 [2024-10-17 17:49:41.860221] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.000 [2024-10-17 17:49:41.869651] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.000 [2024-10-17 17:49:41.870196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.000 [2024-10-17 17:49:41.870229] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.000 [2024-10-17 17:49:41.870237] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.000 [2024-10-17 17:49:41.870402] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.000 [2024-10-17 17:49:41.870553] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.000 [2024-10-17 17:49:41.870559] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.000 [2024-10-17 17:49:41.870564] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.000 [2024-10-17 17:49:41.872969] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.000 [2024-10-17 17:49:41.882265] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.000 [2024-10-17 17:49:41.882731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.000 [2024-10-17 17:49:41.882752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.000 [2024-10-17 17:49:41.882759] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.000 [2024-10-17 17:49:41.882912] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.000 [2024-10-17 17:49:41.883062] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.000 [2024-10-17 17:49:41.883068] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.000 [2024-10-17 17:49:41.883074] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.000 [2024-10-17 17:49:41.885471] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.000 [2024-10-17 17:49:41.894916] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.000 [2024-10-17 17:49:41.895431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.000 [2024-10-17 17:49:41.895462] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.000 [2024-10-17 17:49:41.895470] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.000 [2024-10-17 17:49:41.895634] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.000 [2024-10-17 17:49:41.895794] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.000 [2024-10-17 17:49:41.895801] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.000 [2024-10-17 17:49:41.895807] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.000 [2024-10-17 17:49:41.898206] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.000 [2024-10-17 17:49:41.907519] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.000 [2024-10-17 17:49:41.908097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.000 [2024-10-17 17:49:41.908127] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.000 [2024-10-17 17:49:41.908135] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.000 [2024-10-17 17:49:41.908300] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.000 [2024-10-17 17:49:41.908455] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.000 [2024-10-17 17:49:41.908461] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.000 [2024-10-17 17:49:41.908467] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.000 [2024-10-17 17:49:41.910870] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.263 [2024-10-17 17:49:41.920179] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.263 [2024-10-17 17:49:41.920637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.263 [2024-10-17 17:49:41.920651] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.263 [2024-10-17 17:49:41.920657] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.263 [2024-10-17 17:49:41.920810] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.263 [2024-10-17 17:49:41.920960] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.263 [2024-10-17 17:49:41.920966] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.263 [2024-10-17 17:49:41.920971] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.263 [2024-10-17 17:49:41.923368] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.263 [2024-10-17 17:49:41.932809] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.263 [2024-10-17 17:49:41.933342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.263 [2024-10-17 17:49:41.933373] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.263 [2024-10-17 17:49:41.933382] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.263 [2024-10-17 17:49:41.933546] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.263 [2024-10-17 17:49:41.933704] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.263 [2024-10-17 17:49:41.933711] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.263 [2024-10-17 17:49:41.933716] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.263 [2024-10-17 17:49:41.936113] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.263 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh: line 35: 226047 Killed "${NVMF_APP[@]}" "$@" 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@36 -- # tgt_init 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@724 -- # xtrace_disable 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:34.263 [2024-10-17 17:49:41.945421] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.263 [2024-10-17 17:49:41.945798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.263 [2024-10-17 17:49:41.945829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.263 [2024-10-17 17:49:41.945838] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.263 [2024-10-17 17:49:41.946008] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # nvmfpid=227746 00:28:34.263 [2024-10-17 17:49:41.946160] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.263 [2024-10-17 17:49:41.946167] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.263 [2024-10-17 17:49:41.946173] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # waitforlisten 227746 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@831 -- # '[' -z 227746 ']' 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:34.263 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:34.263 17:49:41 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:34.263 [2024-10-17 17:49:41.948574] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.263 [2024-10-17 17:49:41.958027] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.263 [2024-10-17 17:49:41.958484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.263 [2024-10-17 17:49:41.958498] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.263 [2024-10-17 17:49:41.958504] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.263 [2024-10-17 17:49:41.958653] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.263 [2024-10-17 17:49:41.958806] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.263 [2024-10-17 17:49:41.958813] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.263 [2024-10-17 17:49:41.958818] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.263 [2024-10-17 17:49:41.961210] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.263 [2024-10-17 17:49:41.970652] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.263 [2024-10-17 17:49:41.971154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.263 [2024-10-17 17:49:41.971169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.263 [2024-10-17 17:49:41.971175] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.263 [2024-10-17 17:49:41.971325] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.263 [2024-10-17 17:49:41.971473] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.263 [2024-10-17 17:49:41.971479] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.263 [2024-10-17 17:49:41.971484] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.263 [2024-10-17 17:49:41.973887] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.263 [2024-10-17 17:49:41.983337] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.263 [2024-10-17 17:49:41.983807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.263 [2024-10-17 17:49:41.983837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.263 [2024-10-17 17:49:41.983847] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.263 [2024-10-17 17:49:41.984014] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.263 [2024-10-17 17:49:41.984167] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.263 [2024-10-17 17:49:41.984173] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.263 [2024-10-17 17:49:41.984179] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.263 [2024-10-17 17:49:41.986581] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.263 [2024-10-17 17:49:41.996035] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.263 [2024-10-17 17:49:41.996569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.263 [2024-10-17 17:49:41.996599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.263 [2024-10-17 17:49:41.996608] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.263 [2024-10-17 17:49:41.996779] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.263 [2024-10-17 17:49:41.996931] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.263 [2024-10-17 17:49:41.996937] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:41.996942] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:41.999170] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:34.264 [2024-10-17 17:49:41.999214] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:34.264 [2024-10-17 17:49:41.999342] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.008656] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.009186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.009216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.009226] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.009390] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.009541] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.009548] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.009554] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.011965] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.021280] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.021790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.021821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.021830] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.021997] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.022148] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.022154] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.022160] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.024569] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.033958] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.034420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.034450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.034459] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.034623] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.034780] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.034787] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.034793] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.037193] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.046633] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.047098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.047112] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.047118] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.047267] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.047415] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.047421] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.047426] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.049822] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.059258] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.059715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.059728] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.059738] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.059887] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.060035] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.060041] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.060047] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.062440] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.071882] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.072375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.072406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.072415] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.072579] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.072736] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.072744] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.072749] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.075143] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.083547] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:28:34.264 [2024-10-17 17:49:42.084452] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.085051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.085081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.085090] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.085256] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.085407] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.085414] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.085421] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.087828] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.097155] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.097746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.097777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.097786] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.097954] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.098109] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.264 [2024-10-17 17:49:42.098116] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.264 [2024-10-17 17:49:42.098121] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.264 [2024-10-17 17:49:42.100529] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.264 [2024-10-17 17:49:42.109851] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.264 [2024-10-17 17:49:42.110316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.264 [2024-10-17 17:49:42.110346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.264 [2024-10-17 17:49:42.110356] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.264 [2024-10-17 17:49:42.110521] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.264 [2024-10-17 17:49:42.110672] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.265 [2024-10-17 17:49:42.110679] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.265 [2024-10-17 17:49:42.110684] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.265 [2024-10-17 17:49:42.113091] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.265 [2024-10-17 17:49:42.113172] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:28:34.265 [2024-10-17 17:49:42.113193] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:28:34.265 [2024-10-17 17:49:42.113199] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:28:34.265 [2024-10-17 17:49:42.113205] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:28:34.265 [2024-10-17 17:49:42.113210] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:28:34.265 [2024-10-17 17:49:42.114333] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:28:34.265 [2024-10-17 17:49:42.114490] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:34.265 [2024-10-17 17:49:42.114492] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:28:34.265 [2024-10-17 17:49:42.122549] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.265 [2024-10-17 17:49:42.123222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.265 [2024-10-17 17:49:42.123254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.265 [2024-10-17 17:49:42.123263] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.265 [2024-10-17 17:49:42.123439] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.265 [2024-10-17 17:49:42.123591] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.265 [2024-10-17 17:49:42.123597] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.265 [2024-10-17 17:49:42.123603] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.265 [2024-10-17 17:49:42.126007] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.265 [2024-10-17 17:49:42.135171] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.265 [2024-10-17 17:49:42.135718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.265 [2024-10-17 17:49:42.135749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.265 [2024-10-17 17:49:42.135762] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.265 [2024-10-17 17:49:42.135926] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.265 [2024-10-17 17:49:42.136078] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.265 [2024-10-17 17:49:42.136084] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.265 [2024-10-17 17:49:42.136091] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.265 [2024-10-17 17:49:42.138493] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.265 [2024-10-17 17:49:42.147803] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.265 [2024-10-17 17:49:42.148398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.265 [2024-10-17 17:49:42.148428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.265 [2024-10-17 17:49:42.148437] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.265 [2024-10-17 17:49:42.148602] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.265 [2024-10-17 17:49:42.148760] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.265 [2024-10-17 17:49:42.148768] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.265 [2024-10-17 17:49:42.148773] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.265 [2024-10-17 17:49:42.151169] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.265 [2024-10-17 17:49:42.160479] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.265 [2024-10-17 17:49:42.161073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.265 [2024-10-17 17:49:42.161103] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.265 [2024-10-17 17:49:42.161112] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.265 [2024-10-17 17:49:42.161277] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.265 [2024-10-17 17:49:42.161429] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.265 [2024-10-17 17:49:42.161435] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.265 [2024-10-17 17:49:42.161441] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.265 [2024-10-17 17:49:42.163848] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.265 [2024-10-17 17:49:42.173151] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.265 [2024-10-17 17:49:42.173708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.265 [2024-10-17 17:49:42.173739] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.265 [2024-10-17 17:49:42.173748] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.265 [2024-10-17 17:49:42.173913] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.265 [2024-10-17 17:49:42.174065] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.265 [2024-10-17 17:49:42.174075] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.265 [2024-10-17 17:49:42.174080] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.265 [2024-10-17 17:49:42.176478] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.527 [2024-10-17 17:49:42.185824] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.527 [2024-10-17 17:49:42.186158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.527 [2024-10-17 17:49:42.186173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.527 [2024-10-17 17:49:42.186179] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.527 [2024-10-17 17:49:42.186327] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.527 [2024-10-17 17:49:42.186476] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.527 [2024-10-17 17:49:42.186481] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.527 [2024-10-17 17:49:42.186486] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.527 [2024-10-17 17:49:42.188885] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.527 [2024-10-17 17:49:42.198481] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.527 [2024-10-17 17:49:42.198881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.527 [2024-10-17 17:49:42.198894] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.527 [2024-10-17 17:49:42.198900] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.527 [2024-10-17 17:49:42.199049] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.527 [2024-10-17 17:49:42.199197] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.527 [2024-10-17 17:49:42.199202] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.527 [2024-10-17 17:49:42.199207] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.527 [2024-10-17 17:49:42.201596] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.527 [2024-10-17 17:49:42.211052] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.527 [2024-10-17 17:49:42.211510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.527 [2024-10-17 17:49:42.211522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.527 [2024-10-17 17:49:42.211527] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.527 [2024-10-17 17:49:42.211676] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.527 [2024-10-17 17:49:42.211828] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.527 [2024-10-17 17:49:42.211834] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.527 [2024-10-17 17:49:42.211839] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.527 [2024-10-17 17:49:42.214229] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.527 [2024-10-17 17:49:42.223680] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.527 [2024-10-17 17:49:42.224222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.527 [2024-10-17 17:49:42.224253] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.527 [2024-10-17 17:49:42.224262] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.527 [2024-10-17 17:49:42.224426] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.527 [2024-10-17 17:49:42.224578] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.527 [2024-10-17 17:49:42.224584] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.527 [2024-10-17 17:49:42.224590] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.527 [2024-10-17 17:49:42.226992] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.527 [2024-10-17 17:49:42.236296] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.527 [2024-10-17 17:49:42.236698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.527 [2024-10-17 17:49:42.236729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.527 [2024-10-17 17:49:42.236738] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.527 [2024-10-17 17:49:42.236905] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.237056] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.237063] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.237068] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.239463] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.248913] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.249356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.249385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.249394] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.249560] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.249718] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.249726] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.249731] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.252124] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.261565] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.262153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.262183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.262191] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.262359] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.262511] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.262517] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.262524] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.264925] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.274226] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.274807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.274837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.274846] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.275011] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.275162] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.275168] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.275174] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.277571] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.286872] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.287430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.287460] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.287468] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.287632] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.287791] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.287798] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.287803] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.290200] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.299516] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.299967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.299998] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.300006] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.300173] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.300324] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.300330] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.300339] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.302743] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.312182] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.312776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.312806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.312815] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.312981] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.313133] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.313139] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.313145] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.315546] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.324911] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.325473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.325504] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.325513] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.325678] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.325835] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.325842] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.325848] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.328241] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.337546] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.338127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.338157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.338166] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.338331] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.338482] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.338488] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.338493] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.340894] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.528 [2024-10-17 17:49:42.350193] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.528 [2024-10-17 17:49:42.350755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.528 [2024-10-17 17:49:42.350785] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.528 [2024-10-17 17:49:42.350794] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.528 [2024-10-17 17:49:42.350958] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.528 [2024-10-17 17:49:42.351109] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.528 [2024-10-17 17:49:42.351116] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.528 [2024-10-17 17:49:42.351122] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.528 [2024-10-17 17:49:42.353522] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.529 [2024-10-17 17:49:42.362823] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.529 [2024-10-17 17:49:42.363249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.529 [2024-10-17 17:49:42.363264] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.529 [2024-10-17 17:49:42.363269] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.529 [2024-10-17 17:49:42.363418] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.529 [2024-10-17 17:49:42.363566] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.529 [2024-10-17 17:49:42.363572] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.529 [2024-10-17 17:49:42.363577] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.529 [2024-10-17 17:49:42.365970] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.529 [2024-10-17 17:49:42.375410] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.529 [2024-10-17 17:49:42.375720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.529 [2024-10-17 17:49:42.375733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.529 [2024-10-17 17:49:42.375739] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.529 [2024-10-17 17:49:42.375887] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.529 [2024-10-17 17:49:42.376035] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.529 [2024-10-17 17:49:42.376040] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.529 [2024-10-17 17:49:42.376045] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.529 [2024-10-17 17:49:42.378434] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.529 [2024-10-17 17:49:42.388077] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.529 [2024-10-17 17:49:42.388584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.529 [2024-10-17 17:49:42.388614] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.529 [2024-10-17 17:49:42.388622] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.529 [2024-10-17 17:49:42.388793] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.529 [2024-10-17 17:49:42.388952] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.529 [2024-10-17 17:49:42.388958] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.529 [2024-10-17 17:49:42.388963] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.529 [2024-10-17 17:49:42.391367] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.529 [2024-10-17 17:49:42.400673] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.529 [2024-10-17 17:49:42.401165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.529 [2024-10-17 17:49:42.401195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.529 [2024-10-17 17:49:42.401204] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.529 [2024-10-17 17:49:42.401369] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.529 [2024-10-17 17:49:42.401521] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.529 [2024-10-17 17:49:42.401527] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.529 [2024-10-17 17:49:42.401532] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.529 [2024-10-17 17:49:42.403931] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.529 [2024-10-17 17:49:42.413376] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.529 [2024-10-17 17:49:42.413987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.529 [2024-10-17 17:49:42.414018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.529 [2024-10-17 17:49:42.414027] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.529 [2024-10-17 17:49:42.414191] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.529 [2024-10-17 17:49:42.414342] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.529 [2024-10-17 17:49:42.414349] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.529 [2024-10-17 17:49:42.414354] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.529 4364.00 IOPS, 17.05 MiB/s [2024-10-17T15:49:42.448Z] [2024-10-17 17:49:42.417885] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.529 [2024-10-17 17:49:42.426076] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.529 [2024-10-17 17:49:42.426608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.529 [2024-10-17 17:49:42.426639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.529 [2024-10-17 17:49:42.426647] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.529 [2024-10-17 17:49:42.426818] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.529 [2024-10-17 17:49:42.426969] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.529 [2024-10-17 17:49:42.426976] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.529 [2024-10-17 17:49:42.426985] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.529 [2024-10-17 17:49:42.429379] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.529 [2024-10-17 17:49:42.438676] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.529 [2024-10-17 17:49:42.439269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.529 [2024-10-17 17:49:42.439299] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.529 [2024-10-17 17:49:42.439308] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.529 [2024-10-17 17:49:42.439472] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.529 [2024-10-17 17:49:42.439623] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.529 [2024-10-17 17:49:42.439629] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.529 [2024-10-17 17:49:42.439635] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.529 [2024-10-17 17:49:42.442033] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.451333] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.451550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.451565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.451571] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.451724] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.451874] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.451880] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.792 [2024-10-17 17:49:42.451885] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.792 [2024-10-17 17:49:42.454276] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.464179] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.464641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.464654] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.464659] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.464813] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.464962] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.464968] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.792 [2024-10-17 17:49:42.464973] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.792 [2024-10-17 17:49:42.467363] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.476808] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.477212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.477247] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.477257] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.477422] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.477574] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.477581] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.792 [2024-10-17 17:49:42.477586] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.792 [2024-10-17 17:49:42.479987] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.489438] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.489842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.489873] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.489881] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.490048] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.490199] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.490205] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.792 [2024-10-17 17:49:42.490211] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.792 [2024-10-17 17:49:42.492618] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.502079] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.502638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.502669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.502677] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.502846] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.502998] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.503004] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.792 [2024-10-17 17:49:42.503010] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.792 [2024-10-17 17:49:42.505406] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.514712] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.515149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.515180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.515189] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.515353] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.515508] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.515515] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.792 [2024-10-17 17:49:42.515520] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.792 [2024-10-17 17:49:42.517920] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.527367] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.527839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.527854] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.527860] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.528009] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.528157] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.528163] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.792 [2024-10-17 17:49:42.528168] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.792 [2024-10-17 17:49:42.530561] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.792 [2024-10-17 17:49:42.540000] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.792 [2024-10-17 17:49:42.540415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.792 [2024-10-17 17:49:42.540446] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.792 [2024-10-17 17:49:42.540454] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.792 [2024-10-17 17:49:42.540621] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.792 [2024-10-17 17:49:42.540779] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.792 [2024-10-17 17:49:42.540785] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.540791] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.543186] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.552623] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.552968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.552983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.552989] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.553138] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.553286] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.553292] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.553297] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.555689] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.565269] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.565726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.565739] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.565744] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.565893] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.566041] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.566046] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.566052] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.568441] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.577877] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.578177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.578188] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.578194] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.578342] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.578489] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.578495] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.578500] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.580925] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.590504] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.591039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.591070] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.591079] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.591243] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.591394] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.591401] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.591406] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.593811] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.603140] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.603632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.603647] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.603656] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.603808] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.603957] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.603963] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.603968] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.606356] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.615789] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.616094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.616106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.616111] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.616260] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.616408] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.616414] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.616419] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.618813] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.628400] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.629017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.629048] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.629056] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.629221] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.629372] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.629378] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.629384] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.631788] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.641095] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.641538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.641567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.641576] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.641749] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.641901] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.641910] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.641916] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.644311] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.653763] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.654279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.654310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.654318] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.654482] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.654634] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.654640] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.793 [2024-10-17 17:49:42.654646] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.793 [2024-10-17 17:49:42.657046] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.793 [2024-10-17 17:49:42.666355] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.793 [2024-10-17 17:49:42.666827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.793 [2024-10-17 17:49:42.666857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.793 [2024-10-17 17:49:42.666866] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.793 [2024-10-17 17:49:42.667033] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.793 [2024-10-17 17:49:42.667184] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.793 [2024-10-17 17:49:42.667190] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.794 [2024-10-17 17:49:42.667196] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.794 [2024-10-17 17:49:42.669594] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.794 [2024-10-17 17:49:42.679047] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.794 [2024-10-17 17:49:42.679398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.794 [2024-10-17 17:49:42.679413] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.794 [2024-10-17 17:49:42.679419] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.794 [2024-10-17 17:49:42.679567] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.794 [2024-10-17 17:49:42.679721] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.794 [2024-10-17 17:49:42.679727] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.794 [2024-10-17 17:49:42.679732] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.794 [2024-10-17 17:49:42.682122] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.794 [2024-10-17 17:49:42.691713] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.794 [2024-10-17 17:49:42.692040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.794 [2024-10-17 17:49:42.692053] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.794 [2024-10-17 17:49:42.692058] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.794 [2024-10-17 17:49:42.692206] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.794 [2024-10-17 17:49:42.692356] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.794 [2024-10-17 17:49:42.692362] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.794 [2024-10-17 17:49:42.692367] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.794 [2024-10-17 17:49:42.694762] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:34.794 [2024-10-17 17:49:42.704351] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:34.794 [2024-10-17 17:49:42.704652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:34.794 [2024-10-17 17:49:42.704665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:34.794 [2024-10-17 17:49:42.704671] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:34.794 [2024-10-17 17:49:42.704823] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:34.794 [2024-10-17 17:49:42.704971] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:34.794 [2024-10-17 17:49:42.704977] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:34.794 [2024-10-17 17:49:42.704982] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:34.794 [2024-10-17 17:49:42.707533] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.056 [2024-10-17 17:49:42.716991] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.056 [2024-10-17 17:49:42.717342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.056 [2024-10-17 17:49:42.717355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.056 [2024-10-17 17:49:42.717361] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.056 [2024-10-17 17:49:42.717509] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.056 [2024-10-17 17:49:42.717658] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.056 [2024-10-17 17:49:42.717664] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.056 [2024-10-17 17:49:42.717669] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.056 [2024-10-17 17:49:42.720063] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.056 [2024-10-17 17:49:42.729659] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.056 [2024-10-17 17:49:42.730117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.056 [2024-10-17 17:49:42.730130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.730136] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.730288] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.730436] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.730443] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.730448] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.732840] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 [2024-10-17 17:49:42.742276] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.742735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.742747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.742753] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.742902] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.743051] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.743058] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.743064] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.745455] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 [2024-10-17 17:49:42.754898] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.755354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.755366] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.755372] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.755520] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.755669] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.755675] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.755681] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.758073] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 [2024-10-17 17:49:42.767516] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.767977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.768008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.768018] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.768182] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.768333] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.768341] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.768350] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.770754] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 [2024-10-17 17:49:42.780216] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.780531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.780546] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.780551] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.780704] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.780853] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.780858] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.780863] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.783263] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 [2024-10-17 17:49:42.792855] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.793359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.793389] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.793398] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.793562] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.793719] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.793726] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.793732] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.796129] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # return 0 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@730 -- # xtrace_disable 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:35.057 [2024-10-17 17:49:42.805474] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.805831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.805848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.805853] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.806002] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.806151] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.806158] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.806166] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.808558] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 [2024-10-17 17:49:42.818149] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.818558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.818571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.818576] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.818729] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.818878] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.818884] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.818889] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.821280] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 [2024-10-17 17:49:42.830740] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.831224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.831255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.831264] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.831428] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.831579] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.831586] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.831591] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.833993] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:35.057 [2024-10-17 17:49:42.843312] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.057 [2024-10-17 17:49:42.843663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.057 [2024-10-17 17:49:42.843678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.057 [2024-10-17 17:49:42.843683] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.057 [2024-10-17 17:49:42.843837] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.057 [2024-10-17 17:49:42.843986] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.057 [2024-10-17 17:49:42.843992] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.057 [2024-10-17 17:49:42.844001] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.057 [2024-10-17 17:49:42.846245] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:35.057 [2024-10-17 17:49:42.846393] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.057 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:35.058 [2024-10-17 17:49:42.855978] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.058 [2024-10-17 17:49:42.856296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.058 [2024-10-17 17:49:42.856308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.058 [2024-10-17 17:49:42.856314] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.058 [2024-10-17 17:49:42.856462] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.058 [2024-10-17 17:49:42.856609] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.058 [2024-10-17 17:49:42.856615] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.058 [2024-10-17 17:49:42.856620] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.058 [2024-10-17 17:49:42.859017] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.058 [2024-10-17 17:49:42.868600] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.058 [2024-10-17 17:49:42.868994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.058 [2024-10-17 17:49:42.869007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.058 [2024-10-17 17:49:42.869012] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.058 [2024-10-17 17:49:42.869160] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.058 [2024-10-17 17:49:42.869308] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.058 [2024-10-17 17:49:42.869314] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.058 [2024-10-17 17:49:42.869319] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.058 [2024-10-17 17:49:42.871713] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.058 Malloc0 00:28:35.058 [2024-10-17 17:49:42.881299] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:28:35.058 [2024-10-17 17:49:42.881900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.058 [2024-10-17 17:49:42.881931] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.058 [2024-10-17 17:49:42.881940] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.058 [2024-10-17 17:49:42.882104] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:35.058 [2024-10-17 17:49:42.882259] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.058 [2024-10-17 17:49:42.882267] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.058 [2024-10-17 17:49:42.882272] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:35.058 [2024-10-17 17:49:42.884675] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:35.058 [2024-10-17 17:49:42.894002] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:35.058 [2024-10-17 17:49:42.894472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.058 [2024-10-17 17:49:42.894487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.058 [2024-10-17 17:49:42.894493] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.058 [2024-10-17 17:49:42.894641] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.058 [2024-10-17 17:49:42.894794] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.058 [2024-10-17 17:49:42.894800] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.058 [2024-10-17 17:49:42.894805] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.058 [2024-10-17 17:49:42.897198] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:35.058 [2024-10-17 17:49:42.906649] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.058 [2024-10-17 17:49:42.907177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:35.058 [2024-10-17 17:49:42.907209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x60c700 with addr=10.0.0.2, port=4420 00:28:35.058 [2024-10-17 17:49:42.907217] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x60c700 is same with the state(6) to be set 00:28:35.058 [2024-10-17 17:49:42.907382] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x60c700 (9): Bad file descriptor 00:28:35.058 [2024-10-17 17:49:42.907534] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:28:35.058 [2024-10-17 17:49:42.907540] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:28:35.058 [2024-10-17 17:49:42.907546] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:28:35.058 [2024-10-17 17:49:42.909947] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:28:35.058 [2024-10-17 17:49:42.912478] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:35.058 17:49:42 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@38 -- # wait 226550 00:28:35.058 [2024-10-17 17:49:42.919260] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:28:35.319 [2024-10-17 17:49:42.985822] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:28:36.520 4552.00 IOPS, 17.78 MiB/s [2024-10-17T15:49:45.823Z] 5598.25 IOPS, 21.87 MiB/s [2024-10-17T15:49:46.764Z] 6436.11 IOPS, 25.14 MiB/s [2024-10-17T15:49:47.706Z] 7090.40 IOPS, 27.70 MiB/s [2024-10-17T15:49:48.649Z] 7624.73 IOPS, 29.78 MiB/s [2024-10-17T15:49:49.627Z] 8070.75 IOPS, 31.53 MiB/s [2024-10-17T15:49:50.615Z] 8444.69 IOPS, 32.99 MiB/s [2024-10-17T15:49:51.553Z] 8768.14 IOPS, 34.25 MiB/s [2024-10-17T15:49:51.553Z] 9053.00 IOPS, 35.36 MiB/s 00:28:43.634 Latency(us) 00:28:43.634 [2024-10-17T15:49:51.553Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:43.634 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:28:43.634 Verification LBA range: start 0x0 length 0x4000 00:28:43.634 Nvme1n1 : 15.01 9058.20 35.38 13631.56 0.00 5622.08 549.55 15182.51 00:28:43.634 [2024-10-17T15:49:51.553Z] =================================================================================================================== 00:28:43.634 [2024-10-17T15:49:51.553Z] Total : 9058.20 35.38 13631.56 0.00 5622.08 549.55 15182.51 00:28:43.634 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@39 -- # sync 00:28:43.634 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:28:43.634 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:43.634 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@42 -- # trap - SIGINT SIGTERM EXIT 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@44 -- # nvmftestfini 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@514 -- # nvmfcleanup 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@121 -- # sync 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@124 -- # set +e 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@125 -- # for i in {1..20} 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:28:43.894 rmmod nvme_tcp 00:28:43.894 rmmod nvme_fabrics 00:28:43.894 rmmod nvme_keyring 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@128 -- # set -e 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@129 -- # return 0 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@515 -- # '[' -n 227746 ']' 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@516 -- # killprocess 227746 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@950 -- # '[' -z 227746 ']' 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@954 -- # kill -0 227746 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@955 -- # uname 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 227746 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@968 -- # echo 'killing process with pid 227746' 00:28:43.894 killing process with pid 227746 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@969 -- # kill 227746 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@974 -- # wait 227746 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@297 -- # iptr 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@789 -- # iptables-save 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:28:43.894 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@789 -- # iptables-restore 00:28:44.154 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:28:44.154 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:28:44.155 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:44.155 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:44.155 17:49:51 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:28:46.067 00:28:46.067 real 0m28.360s 00:28:46.067 user 1m3.401s 00:28:46.067 sys 0m7.679s 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:28:46.067 ************************************ 00:28:46.067 END TEST nvmf_bdevperf 00:28:46.067 ************************************ 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@48 -- # run_test nvmf_target_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:28:46.067 ************************************ 00:28:46.067 START TEST nvmf_target_disconnect 00:28:46.067 ************************************ 00:28:46.067 17:49:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:28:46.329 * Looking for test storage... 00:28:46.329 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1691 -- # lcov --version 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@345 -- # : 1 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # decimal 1 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=1 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 1 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # decimal 2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # return 0 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:28:46.329 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:46.329 --rc genhtml_branch_coverage=1 00:28:46.329 --rc genhtml_function_coverage=1 00:28:46.329 --rc genhtml_legend=1 00:28:46.329 --rc geninfo_all_blocks=1 00:28:46.329 --rc geninfo_unexecuted_blocks=1 00:28:46.329 00:28:46.329 ' 00:28:46.329 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:28:46.329 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:46.329 --rc genhtml_branch_coverage=1 00:28:46.329 --rc genhtml_function_coverage=1 00:28:46.329 --rc genhtml_legend=1 00:28:46.329 --rc geninfo_all_blocks=1 00:28:46.329 --rc geninfo_unexecuted_blocks=1 00:28:46.329 00:28:46.330 ' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:28:46.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:46.330 --rc genhtml_branch_coverage=1 00:28:46.330 --rc genhtml_function_coverage=1 00:28:46.330 --rc genhtml_legend=1 00:28:46.330 --rc geninfo_all_blocks=1 00:28:46.330 --rc geninfo_unexecuted_blocks=1 00:28:46.330 00:28:46.330 ' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:28:46.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:46.330 --rc genhtml_branch_coverage=1 00:28:46.330 --rc genhtml_function_coverage=1 00:28:46.330 --rc genhtml_legend=1 00:28:46.330 --rc geninfo_all_blocks=1 00:28:46.330 --rc geninfo_unexecuted_blocks=1 00:28:46.330 00:28:46.330 ' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # uname -s 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@5 -- # export PATH 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@51 -- # : 0 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:28:46.330 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@11 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@13 -- # MALLOC_BDEV_SIZE=64 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@69 -- # nvmftestinit 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@474 -- # prepare_net_devs 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@436 -- # local -g is_hw=no 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@438 -- # remove_spdk_ns 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:28:46.330 17:49:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # e810=() 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # x722=() 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:28:54.475 Found 0000:31:00.0 (0x8086 - 0x159b) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:28:54.475 Found 0000:31:00.1 (0x8086 - 0x159b) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:28:54.475 Found net devices under 0000:31:00.0: cvl_0_0 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:28:54.475 Found net devices under 0000:31:00.1: cvl_0_1 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # is_hw=yes 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:28:54.475 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:28:54.476 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:28:54.476 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.631 ms 00:28:54.476 00:28:54.476 --- 10.0.0.2 ping statistics --- 00:28:54.476 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:54.476 rtt min/avg/max/mdev = 0.631/0.631/0.631/0.000 ms 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:28:54.476 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:28:54.476 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.295 ms 00:28:54.476 00:28:54.476 --- 10.0.0.1 ping statistics --- 00:28:54.476 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:54.476 rtt min/avg/max/mdev = 0.295/0.295/0.295/0.000 ms 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@448 -- # return 0 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@70 -- # run_test nvmf_target_disconnect_tc1 nvmf_target_disconnect_tc1 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:28:54.476 ************************************ 00:28:54.476 START TEST nvmf_target_disconnect_tc1 00:28:54.476 ************************************ 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1125 -- # nvmf_target_disconnect_tc1 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- host/target_disconnect.sh@32 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@650 -- # local es=0 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect ]] 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:28:54.476 [2024-10-17 17:50:01.980304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:54.476 [2024-10-17 17:50:01.980400] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1a77020 with addr=10.0.0.2, port=4420 00:28:54.476 [2024-10-17 17:50:01.980434] nvme_tcp.c:2723:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:28:54.476 [2024-10-17 17:50:01.980447] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:28:54.476 [2024-10-17 17:50:01.980456] nvme.c: 939:spdk_nvme_probe_ext: *ERROR*: Create probe context failed 00:28:54.476 spdk_nvme_probe() failed for transport address '10.0.0.2' 00:28:54.476 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect: errors occurred 00:28:54.476 Initializing NVMe Controllers 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@653 -- # es=1 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:28:54.476 00:28:54.476 real 0m0.132s 00:28:54.476 user 0m0.050s 00:28:54.476 sys 0m0.081s 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:54.476 17:50:01 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@10 -- # set +x 00:28:54.476 ************************************ 00:28:54.476 END TEST nvmf_target_disconnect_tc1 00:28:54.476 ************************************ 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@71 -- # run_test nvmf_target_disconnect_tc2 nvmf_target_disconnect_tc2 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:28:54.476 ************************************ 00:28:54.476 START TEST nvmf_target_disconnect_tc2 00:28:54.476 ************************************ 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1125 -- # nvmf_target_disconnect_tc2 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@37 -- # disconnect_init 10.0.0.2 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # nvmfpid=233848 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # waitforlisten 233848 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@831 -- # '[' -z 233848 ']' 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:54.476 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:54.476 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:54.477 [2024-10-17 17:50:02.143682] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:54.477 [2024-10-17 17:50:02.143747] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:54.477 [2024-10-17 17:50:02.234465] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:28:54.477 [2024-10-17 17:50:02.288117] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:28:54.477 [2024-10-17 17:50:02.288164] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:28:54.477 [2024-10-17 17:50:02.288173] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:28:54.477 [2024-10-17 17:50:02.288180] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:28:54.477 [2024-10-17 17:50:02.288187] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:28:54.477 [2024-10-17 17:50:02.290276] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:28:54.477 [2024-10-17 17:50:02.290441] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:28:54.477 [2024-10-17 17:50:02.290599] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:28:54.477 [2024-10-17 17:50:02.290600] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:28:55.421 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:55.421 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # return 0 00:28:55.421 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:28:55.421 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:28:55.421 17:50:02 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:55.421 Malloc0 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:55.421 [2024-10-17 17:50:03.065430] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:55.421 [2024-10-17 17:50:03.105861] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:55.421 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:55.422 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@42 -- # reconnectpid=234201 00:28:55.422 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@44 -- # sleep 2 00:28:55.422 17:50:03 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:28:57.343 17:50:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@45 -- # kill -9 233848 00:28:57.344 17:50:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@47 -- # sleep 2 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Write completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 Read completed with error (sct=0, sc=8) 00:28:57.344 starting I/O failed 00:28:57.344 [2024-10-17 17:50:05.142568] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:28:57.344 [2024-10-17 17:50:05.143154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.143226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.143651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.143665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.144105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.144154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.144508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.144529] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.144924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.144972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.145309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.145323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.145672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.145683] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.146156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.146205] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.146561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.146575] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.147003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.147052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.147407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.147421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.147751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.147763] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.147984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.147995] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.148289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.148300] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.148649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.148660] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.148889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.148900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.149176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.149187] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.149500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.149512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.344 qpair failed and we were unable to recover it. 00:28:57.344 [2024-10-17 17:50:05.149716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.344 [2024-10-17 17:50:05.149728] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.150016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.150027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.150345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.150356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.150591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.150602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.150819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.150830] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.151113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.151124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.151316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.151331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.151663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.151674] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.152045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.152057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.152409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.152422] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.152726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.152737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.152960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.152972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.153265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.153277] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.153609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.153620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.154022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.154033] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.154227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.154239] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.154586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.154597] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.154916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.154928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.155265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.155276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.155592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.155605] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.155865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.155876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.156183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.156194] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.156438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.156450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.156759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.156771] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.157097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.157108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.157487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.157503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.157822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.157834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.158158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.158169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.158420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.158431] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.158735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.158747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.158921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.158932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.159243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.159254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.159564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.159575] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.159872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.159882] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.160282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.345 [2024-10-17 17:50:05.160293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.345 qpair failed and we were unable to recover it. 00:28:57.345 [2024-10-17 17:50:05.160601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.160613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.160856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.160868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.161129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.161140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.161454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.161465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.161806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.161816] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.162149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.162160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.162488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.162498] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.162826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.162836] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.163221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.163231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.163557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.163567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.163912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.163923] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.164241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.164251] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.164553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.164564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.164876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.164888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.165138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.165150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.165450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.165461] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.165775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.165787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.166012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.166022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.166189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.166200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.166553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.166565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.166866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.166878] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.167224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.167235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.167418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.167429] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.167778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.167789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.168113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.168125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.168460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.168471] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.168772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.168783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.169099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.169110] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.169304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.169316] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.169617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.169629] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.169943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.169955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.170278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.170288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.170678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.170688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.171012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.171023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.346 qpair failed and we were unable to recover it. 00:28:57.346 [2024-10-17 17:50:05.171340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.346 [2024-10-17 17:50:05.171350] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.171644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.171654] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.171962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.171973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.172326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.172338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.172629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.172643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.172951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.172964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.173262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.173275] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.173612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.173624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.173945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.173957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.174171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.174182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.174504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.174516] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.174944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.174959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.175257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.175270] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.175498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.175512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.175857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.175872] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.176217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.176231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.176544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.176558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.176732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.176749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.176966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.176979] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.177287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.177302] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.177627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.177641] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.177822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.177837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.178203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.178217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.178541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.178559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.178864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.178879] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.179201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.179215] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.179547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.179561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.179897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.179911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.180238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.180253] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.180575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.180589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.180921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.180936] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.181247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.181261] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.181581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.181595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.181922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.181936] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.182248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.182263] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.182578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.182592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.182986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.183002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.347 [2024-10-17 17:50:05.183293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.347 [2024-10-17 17:50:05.183307] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.347 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.183656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.183670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.184050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.184064] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.184240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.184255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.184547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.184562] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.184889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.184903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.185194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.185207] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.185526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.185539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.185773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.185787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.186100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.186113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.186439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.186453] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.186675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.186689] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.186902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.186922] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.187253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.187271] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.187613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.187631] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.187978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.187996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.188331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.188349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.188676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.188697] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.189023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.189040] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.189367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.189385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.189715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.189734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.190045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.190062] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.190366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.190384] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.190575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.190592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.190813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.190833] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.191168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.191185] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.191522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.191544] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.191920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.191939] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.192286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.192305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.192608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.192627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.192998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.193016] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.193328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.193345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.193682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.193709] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.193983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.348 [2024-10-17 17:50:05.194001] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.348 qpair failed and we were unable to recover it. 00:28:57.348 [2024-10-17 17:50:05.194287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.194305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.194630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.194648] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.194984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.195004] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.195338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.195355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.195698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.195720] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.196005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.196022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.196340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.196358] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.196663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.196680] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.196928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.196947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.197280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.197298] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.197616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.197633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.197860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.197879] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.198220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.198237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.198564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.198588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.198922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.198949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.199322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.199347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.199588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.199615] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.199980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.200007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.200379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.200405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.200771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.200798] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.201155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.201180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.201540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.201566] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.201927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.201953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.202199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.202227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.202585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.202610] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.202979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.203006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.203445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.203470] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.203797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.203824] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.349 qpair failed and we were unable to recover it. 00:28:57.349 [2024-10-17 17:50:05.204192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.349 [2024-10-17 17:50:05.204217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.204584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.204609] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.204947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.204973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.205331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.205357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.205721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.205760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.206087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.206112] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.206470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.206495] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.206725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.206753] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.207137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.207163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.207408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.207433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.207804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.207830] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.208201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.208227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.208645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.208671] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.209026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.209052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.209395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.209421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.209836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.209862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.210219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.210244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.210602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.210630] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.210880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.210914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.211278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.211306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.211662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.211701] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.212069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.212097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.212430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.212458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.212830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.212861] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.213084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.213114] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.213494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.213522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.213881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.213911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.214258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.214286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.214651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.214679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.215001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.215029] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.215405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.215433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.215678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.215716] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.216069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.216097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.216463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.216490] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.216853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.216883] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.217241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.217269] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.217625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.217653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.218017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.350 [2024-10-17 17:50:05.218047] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.350 qpair failed and we were unable to recover it. 00:28:57.350 [2024-10-17 17:50:05.218414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.218443] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.218806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.218835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.219082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.219111] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.219356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.219385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.219758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.219787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.220129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.220159] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.220532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.220565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.220918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.220949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.221327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.221355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.221720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.221749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.222128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.222157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.222515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.222543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.222916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.222945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.223298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.223325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.223559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.223589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.223951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.223981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.224338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.224367] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.224736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.224768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.225157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.225185] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.225542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.225570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.225929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.225959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.226316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.226344] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.226711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.226740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.227111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.227139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.227506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.227535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.227920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.227949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.228310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.228338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.228705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.228734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.229051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.229080] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.229342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.229369] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.229714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.229744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.230104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.230131] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.230492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.230519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.230891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.230921] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.231292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.231320] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.231702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.231732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.232070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.351 [2024-10-17 17:50:05.232099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.351 qpair failed and we were unable to recover it. 00:28:57.351 [2024-10-17 17:50:05.232400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.232428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.232708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.232738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.233016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.233044] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.233421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.233448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.233818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.233849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.234278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.234306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.234613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.234641] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.234984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.235013] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.235372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.235400] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.235832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.235868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.236222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.236249] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.236584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.236612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.236924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.236953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.237299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.237327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.237722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.237752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.238094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.238129] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.238497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.238526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.238898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.238928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.239267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.239295] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.239674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.239712] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.240073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.240101] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.240463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.240492] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.240856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.240887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.241252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.241280] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.241646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.241675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.242017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.242047] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.242446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.242474] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.242884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.242914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.243264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.243293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.243667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.243703] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.244053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.244082] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.244448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.244477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.244744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.244773] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.245146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.245174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.245539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.245567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.352 [2024-10-17 17:50:05.245940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.352 [2024-10-17 17:50:05.245969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.352 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.246326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.246355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.246714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.246745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.247145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.247173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.247526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.247554] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.247893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.247921] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.248276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.248305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.248674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.248712] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.249077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.249104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.249438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.249467] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.249801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.249831] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.250189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.250217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.250586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.250614] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.251007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.251037] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.251407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.251440] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.251769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.251799] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.252172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.252200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.252539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.252567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.252843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.252874] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.253244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.253272] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.253641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.253670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.253950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.253980] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.254351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.254379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.254746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.254776] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.255191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.255220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.255579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.255606] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.255955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.353 [2024-10-17 17:50:05.255985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.353 qpair failed and we were unable to recover it. 00:28:57.353 [2024-10-17 17:50:05.256352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.256380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.256741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.256771] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.257130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.257158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.257534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.257564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.257926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.257955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.258314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.258342] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.258711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.258741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.259117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.259145] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.259510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.259538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.259881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.627 [2024-10-17 17:50:05.259910] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.627 qpair failed and we were unable to recover it. 00:28:57.627 [2024-10-17 17:50:05.260278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.260308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.260657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.260685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.261098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.261128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.261492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.261521] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.261879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.261911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.262163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.262193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.262557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.262586] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.262950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.262980] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.263378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.263406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.263747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.263778] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.264207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.264236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.264598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.264626] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.265009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.265039] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.265404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.265433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.265777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.265806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.266183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.266212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.266607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.266635] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.266856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.266894] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.267246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.267274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.267638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.267666] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.268009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.268038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.268400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.268428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.268797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.268827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.269202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.269230] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.269591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.269620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.269885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.269914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.270354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.270383] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.270747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.270775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.271142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.271170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.271542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.271570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.271828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.271857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.272225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.272254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.272622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.272650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.273025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.273055] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.273419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.273448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.273812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.273842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.274097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.628 [2024-10-17 17:50:05.274125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.628 qpair failed and we were unable to recover it. 00:28:57.628 [2024-10-17 17:50:05.274474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.274502] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.274872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.274902] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.275264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.275292] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.275550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.275582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.275989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.276019] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.276266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.276298] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.276671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.276722] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.277124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.277153] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.277519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.277547] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.277929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.277959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.278321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.278349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.278711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.278740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.279107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.279136] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.279543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.279571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.279943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.279972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.280224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.280252] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.280630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.280657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.281027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.281057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.281301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.281333] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.281598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.281628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.281982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.282026] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.282403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.282433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.282808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.282837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.283208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.283236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.283600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.283628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.284005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.284034] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.284408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.284436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.284804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.284834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.285198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.285226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.285594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.285621] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.285967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.285999] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.286243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.286274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.286628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.286664] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.287015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.287045] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.287270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.287301] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.287652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.287679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.629 qpair failed and we were unable to recover it. 00:28:57.629 [2024-10-17 17:50:05.288071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.629 [2024-10-17 17:50:05.288100] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.288463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.288491] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.288860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.288893] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.289251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.289279] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.289641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.289670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.290024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.290054] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.290419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.290448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.290708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.290742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.291124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.291152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.291518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.291546] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.291930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.291959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.292215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.292245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.292589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.292618] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.293028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.293058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.293427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.293456] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.293821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.293850] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.294213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.294241] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.294607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.294634] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.295004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.295032] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.295389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.295424] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.295847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.295876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.296133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.296160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.296512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.296539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.296793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.296822] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.297264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.297298] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.297642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.297672] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.298034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.298065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.298427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.298457] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.298820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.298849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.299260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.299289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.299650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.299678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.300035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.300063] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.300426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.300454] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.300822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.300853] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.301222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.301250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.630 qpair failed and we were unable to recover it. 00:28:57.630 [2024-10-17 17:50:05.301626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.630 [2024-10-17 17:50:05.301654] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.301911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.301944] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.302315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.302343] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.302728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.302759] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.303111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.303139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.303552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.303579] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.303950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.303981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.304389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.304417] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.304785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.304815] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.305192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.305220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.305592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.305621] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.305886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.305919] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.306262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.306290] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.306548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.306576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.306924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.306953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.307320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.307349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.307596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.307628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.307885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.307917] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.308140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.308173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.308534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.308563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.308929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.308959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.309320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.309348] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.309595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.309626] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.309994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.310024] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.310399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.310427] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.310792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.310821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.311198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.311227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.311586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.311614] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.311980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.312008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.312369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.312404] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.312780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.312810] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.313213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.313242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.313638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.313667] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.314031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.314062] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.314400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.314428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.314795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.314826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.315202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.315230] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.315611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.315638] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.316010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.316041] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.316297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.631 [2024-10-17 17:50:05.316328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.631 qpair failed and we were unable to recover it. 00:28:57.631 [2024-10-17 17:50:05.316669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.316723] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.317057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.317084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.317416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.317445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.317819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.317850] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.318260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.318287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.318641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.318669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.319077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.319106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.319470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.319498] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.319872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.319901] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.320136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.320167] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.320523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.320551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.320919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.320950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.321186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.321213] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.321579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.321607] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.322048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.322078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.322415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.322444] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.322805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.322835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.323197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.323225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.323605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.323632] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.323893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.323926] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.324299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.324327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.324704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.324734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.325095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.325123] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.325372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.325405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.325803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.325833] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.326201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.326230] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.326592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.326620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.326985] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.327014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.632 [2024-10-17 17:50:05.327381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.632 [2024-10-17 17:50:05.327410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.632 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.327776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.327811] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.328177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.328206] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.328572] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.328601] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.328828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.328860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.329204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.329232] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.329615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.329643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.329889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.329923] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.330293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.330322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.330711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.330741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.331100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.331127] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.331490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.331519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.331880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.331911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.332188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.332216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.332599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.332627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.332929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.332959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.333302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.333331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.333700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.333731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.334086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.334114] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.334480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.334508] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.334865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.334893] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.335267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.335295] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.335547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.335578] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.335948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.335978] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.336340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.336369] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.336738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.336767] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.337128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.337156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.337523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.337551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.337924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.337956] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.338202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.338235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.338581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.338609] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.339042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.339073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.339414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.339452] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.339796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.339826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.340164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.340202] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.340536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.340565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.340927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.340957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.341317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.341346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.341519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.341551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.341923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.633 [2024-10-17 17:50:05.341952] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.633 qpair failed and we were unable to recover it. 00:28:57.633 [2024-10-17 17:50:05.342320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.342348] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.342798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.342835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.343233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.343262] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.343598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.343634] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.344047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.344078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.344442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.344471] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.344838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.344869] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.345129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.345158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.345454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.345483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.345857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.345888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.346257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.346287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.346642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.346670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.346837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.346869] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.347251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.347282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.347621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.347658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.348044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.348074] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.348335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.348367] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.348746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.348777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.349136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.349165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.349525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.349553] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.349965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.349995] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.350360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.350388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.350748] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.350777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.351037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.351065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.351431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.351459] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.351832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.351862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.352235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.352263] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.352628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.352658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.353088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.353119] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.353481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.353510] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.353872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.353903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.354310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.354338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.354706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.634 [2024-10-17 17:50:05.354736] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.634 qpair failed and we were unable to recover it. 00:28:57.634 [2024-10-17 17:50:05.355102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.355130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.355357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.355387] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.355766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.355796] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.356170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.356198] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.356553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.356581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.356945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.356974] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.357384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.357412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.357763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.357792] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.358179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.358207] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.358577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.358607] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.358959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.358988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.359325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.359354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.359711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.359742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.360106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.360134] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.360514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.360542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.360959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.360991] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.361332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.361362] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.361708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.361738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.362083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.362111] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.362469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.362497] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.362733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.362764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.363167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.363195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.363566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.363595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.363858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.363887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.364242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.364270] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.364614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.364644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.364996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.365026] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.365394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.365423] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.365785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.365815] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.366185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.366214] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.366566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.366593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.366950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.366980] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.367231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.367259] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.367657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.367685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.368019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.368049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.635 [2024-10-17 17:50:05.368416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.635 [2024-10-17 17:50:05.368451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.635 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.368805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.368835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.369219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.369247] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.369613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.369640] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.370008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.370038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.370397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.370425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.370794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.370823] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.371198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.371226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.371593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.371622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.371973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.372002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.372400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.372428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.372791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.372820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.373187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.373215] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.373584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.373612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.373978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.374008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.374367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.374395] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.374759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.374790] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.375161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.375190] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.375566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.375594] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.375957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.375987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.376361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.376389] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.376753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.376784] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.377058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.377085] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.377423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.377450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.377814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.377843] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.378219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.378249] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.378684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.378724] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.379084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.379113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.379475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.379503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.379879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.379911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.380319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.380347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.380731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.380762] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.381160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.381188] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.381600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.381628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.381995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.382025] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.382384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.382411] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.382775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.382805] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.383169] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.383198] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.383480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.383508] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.636 qpair failed and we were unable to recover it. 00:28:57.636 [2024-10-17 17:50:05.383858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.636 [2024-10-17 17:50:05.383888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.384252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.384286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.384648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.384677] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.385127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.385156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.385514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.385543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.385910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.385939] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.386382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.386410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.386655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.386682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.387065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.387093] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.387462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.387489] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.387719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.387749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.388124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.388152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.388500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.388530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.388863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.388892] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.389253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.389281] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.389644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.389673] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.389991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.390022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.390404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.390432] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.390797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.390827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.391195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.391223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.391625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.391653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.392011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.392042] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.392416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.392443] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.392810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.392841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.393206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.393234] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.393601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.393629] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.393988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.394019] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.637 [2024-10-17 17:50:05.394382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.637 [2024-10-17 17:50:05.394411] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.637 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.394794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.394824] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.395195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.395223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.395593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.395622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.395955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.395984] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.396344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.396373] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.396736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.396768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.397148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.397176] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.397542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.397572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.397941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.397971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.398338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.398368] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.398611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.398643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.399011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.399041] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.399286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.399315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.399667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.399713] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.400070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.400099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.400458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.400488] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.400852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.400883] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.401251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.401281] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.401625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.401653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.401947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.401977] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.402330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.402358] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.402711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.402741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.403095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.403125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.403486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.403516] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.403888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.403918] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.404287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.404317] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.404681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.404725] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.405111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.405140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.405506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.405535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.405899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.405931] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.406128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.406156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.406543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.406570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.406937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.406968] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.407326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.407355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.407718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.407748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.408109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.408136] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.638 [2024-10-17 17:50:05.408503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.638 [2024-10-17 17:50:05.408533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.638 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.408903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.408933] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.409295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.409323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.409690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.409732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.410118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.410148] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.410513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.410541] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.410898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.410930] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.411315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.411344] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.411580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.411610] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.411966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.411997] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.412368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.412396] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.412663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.412704] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.413110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.413140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.413506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.413535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.413849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.413879] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.414257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.414285] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.414648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.414677] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.415040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.415077] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.415426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.415455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.415897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.415927] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.416284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.416313] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.416679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.416720] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.417077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.417105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.417466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.417495] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.417858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.417887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.418248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.418278] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.418645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.418674] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.419059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.419089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.419469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.419497] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.419858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.419889] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.420263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.420292] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.420701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.420733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.421115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.421146] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.421512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.421540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.421906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.421937] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.422300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.422328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.422702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.422733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.639 [2024-10-17 17:50:05.422946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.639 [2024-10-17 17:50:05.422977] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.639 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.423332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.423361] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.423766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.423797] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.424172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.424202] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.424557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.424586] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.425019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.425049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.425293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.425322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.425683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.425725] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.426096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.426125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.426571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.426600] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.426928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.426957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.427323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.427356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.427716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.427745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.428109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.428137] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.428500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.428531] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.428901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.428932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.429300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.429329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.429737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.429767] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.430111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.430140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.430500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.430529] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.430901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.430938] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.431379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.431410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.431733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.431764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.432017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.432050] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.432346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.432374] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.432746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.432777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.433143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.433172] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.433537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.433566] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.433930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.433960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.434327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.434357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.434730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.434760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.435129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.435158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.435511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.435539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.435919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.435950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.436341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.436371] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.436733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.436764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.437143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.437171] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.437535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.437565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.640 [2024-10-17 17:50:05.437938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.640 [2024-10-17 17:50:05.437967] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.640 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.438329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.438358] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.438712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.438741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.439107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.439137] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.439504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.439533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.439899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.439930] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.440289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.440317] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.440677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.440716] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.441070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.441098] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.441466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.441497] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.441857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.441888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.442246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.442275] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.442638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.442668] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.443046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.443079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.443437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.443466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.443840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.443870] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.444231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.444263] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.444655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.444683] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.445044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.445074] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.445304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.445334] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.445658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.445687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.446064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.446095] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.446453] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.446489] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.446857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.446887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.447249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.447281] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.447646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.447675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.448047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.448078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.448454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.448481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.448847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.448878] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.449239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.449270] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.449629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.449658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.450031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.450062] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.450429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.450458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.450813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.450842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.641 qpair failed and we were unable to recover it. 00:28:57.641 [2024-10-17 17:50:05.451206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.641 [2024-10-17 17:50:05.451237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.451591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.451620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.451982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.452012] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.452368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.452399] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.452766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.452796] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.453172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.453200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.453490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.453521] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.453877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.453906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.454263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.454293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.454651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.454681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.455055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.455084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.455453] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.455484] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.455855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.455885] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.456254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.456284] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.456628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.456657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.457021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.457051] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.457292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.457325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.457706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.457737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.458090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.458118] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.458486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.458515] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.458882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.458914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.459276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.459307] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.459683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.459731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.460116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.460146] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.460514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.460544] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.460915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.460945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.461308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.461338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.461709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.461741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.462071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.462105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.462462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.462490] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.462847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.462878] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.642 qpair failed and we were unable to recover it. 00:28:57.642 [2024-10-17 17:50:05.463249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.642 [2024-10-17 17:50:05.463278] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.463642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.463670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.463922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.463955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.464181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.464209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.464456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.464487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.464841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.464870] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.465256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.465285] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.465523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.465550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.465974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.466004] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.466364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.466393] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.466768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.466798] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.466960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.466991] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.467373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.467402] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.467769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.467799] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.468024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.468054] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.468412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.468440] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.468809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.468839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.469174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.469202] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.469575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.469603] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.469968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.469998] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.470349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.470378] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.470594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.470624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.470984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.471014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.471282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.471310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.471711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.471742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.471984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.472016] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.472440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.472469] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.472838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.472868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.473114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.473144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.473498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.473528] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.473885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.473914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.474275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.474304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.474667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.474706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.643 qpair failed and we were unable to recover it. 00:28:57.643 [2024-10-17 17:50:05.475004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.643 [2024-10-17 17:50:05.475032] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.475382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.475409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.475785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.475817] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.476184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.476212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.476579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.476615] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.476853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.476885] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.477103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.477133] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.477490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.477519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.477902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.477932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.478294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.478322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.478615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.478643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.479014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.479044] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.479421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.479448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.479753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.479783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.480141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.480169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.480531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.480559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.480807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.480836] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.481198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.481225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.481587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.481615] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.481920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.481949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.482371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.482399] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.482735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.482765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.483029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.483060] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.483422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.483451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.483805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.483836] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.484209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.484239] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.484594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.484622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.484881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.484910] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.485282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.485311] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.485541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.485571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.485900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.485930] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.486297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.486326] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.486583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.486613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.486985] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.487015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.487360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.487389] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.487752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.487782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.644 [2024-10-17 17:50:05.488154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.644 [2024-10-17 17:50:05.488183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.644 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.488428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.488460] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.488798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.488827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.489236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.489264] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.489498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.489526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.489792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.489821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.490269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.490297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.490659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.490687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.491044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.491079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.491450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.491478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.491761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.491790] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.492041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.492069] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.492413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.492441] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.492802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.492833] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.493198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.493226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.493624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.493652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.494010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.494040] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.494417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.494446] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.494681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.494722] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.495078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.495108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.495475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.495504] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.495863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.495893] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.496249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.496278] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.496648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.496676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.497042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.497071] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.497432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.497459] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.497818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.497848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.498289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.498317] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.498661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.498690] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.499103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.499132] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.499484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.499513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.499805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.499834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.500090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.500117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.500452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.500479] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.500846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.500877] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.501237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.501265] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.645 qpair failed and we were unable to recover it. 00:28:57.645 [2024-10-17 17:50:05.501670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.645 [2024-10-17 17:50:05.501709] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.502053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.502089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.502331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.502362] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.502649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.502676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.503084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.503113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.503479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.503507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.503867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.503897] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.504258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.504287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.504656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.504684] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.505055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.505083] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.505445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.505473] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.505865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.505896] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.506271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.506306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.506658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.506686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.507060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.507089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.507494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.507522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.507887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.507918] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.508277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.508305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.508670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.508707] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.509077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.509104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.509472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.509500] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.509855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.509884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.510271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.510299] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.510567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.510593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.510955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.510984] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.511226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.511258] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.511619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.511649] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.512038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.512068] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.512424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.512453] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.512836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.512867] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.513231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.513259] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.513623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.513651] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.514017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.514046] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.514402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.514429] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.646 [2024-10-17 17:50:05.514802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.646 [2024-10-17 17:50:05.514833] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.646 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.515275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.515303] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.515672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.515710] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.516066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.516094] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.516351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.516379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.516651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.516679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.517076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.517105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.517482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.517510] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.517872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.517902] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.518261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.518289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.518650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.518678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.519001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.519030] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.519283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.519314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.519717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.519747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.520140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.520169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.520530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.520559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.520928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.520958] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.521269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.521299] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.521677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.521726] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.522106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.522135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.522483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.522511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.522885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.522917] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.523286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.523314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.523674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.523712] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.524068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.524095] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.524504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.524532] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.524895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.524924] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.525280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.525308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.525674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.525712] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.525974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.526001] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.526385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.526413] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.526781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.526812] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.527180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.527209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.527563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.527592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.527962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.527992] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.528225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.528252] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.528619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.528648] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.647 [2024-10-17 17:50:05.529070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.647 [2024-10-17 17:50:05.529099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.647 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.529439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.529468] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.529838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.529867] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.530242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.530270] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.530639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.530667] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.530877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.530911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.531287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.531315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.531677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.531720] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.532102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.532132] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.532491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.532519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.532877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.532907] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.648 [2024-10-17 17:50:05.533273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.648 [2024-10-17 17:50:05.533302] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.648 qpair failed and we were unable to recover it. 00:28:57.922 [2024-10-17 17:50:05.533594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.922 [2024-10-17 17:50:05.533624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.922 qpair failed and we were unable to recover it. 00:28:57.922 [2024-10-17 17:50:05.534009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.534040] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.534399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.534427] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.534790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.534820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.535184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.535212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.535584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.535613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.535956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.535995] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.536334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.536363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.536727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.536759] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.537129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.537165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.537306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.537336] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.537802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.537832] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.538200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.538228] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.538592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.538619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.538990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.539020] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.539231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.539259] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.539616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.539644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.540007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.540037] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.540471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.540500] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.540857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.540888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.541251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.541280] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.541648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.541676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.541914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.541943] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.542328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.542357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.542722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.542753] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.542994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.543022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.543371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.543401] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.543782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.543813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.544062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.544092] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.544464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.544493] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.544856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.544888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.545249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.545277] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.545426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.545455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.545724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.545756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.546144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.546172] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.546530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.546559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.546825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.546857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.547203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.547233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.547594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.923 [2024-10-17 17:50:05.547623] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.923 qpair failed and we were unable to recover it. 00:28:57.923 [2024-10-17 17:50:05.547906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.547939] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.548191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.548220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.548600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.548630] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.549037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.549069] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.549427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.549459] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.549668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.549706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.550077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.550105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.550313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.550342] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.550705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.550735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.551081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.551116] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.551467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.551496] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.551858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.551889] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.552257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.552286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.552648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.552676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.553045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.553073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.553444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.553472] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.553840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.553870] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.554233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.554261] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.554622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.554650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.555056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.555086] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.555395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.555423] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.555804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.555834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.556195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.556224] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.556592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.556620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.556986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.557017] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.557374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.557403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.557758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.557788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.558206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.558235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.558603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.558631] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.559004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.559033] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.559397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.559424] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.559786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.559815] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.560128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.924 [2024-10-17 17:50:05.560156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.924 qpair failed and we were unable to recover it. 00:28:57.924 [2024-10-17 17:50:05.560496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.560525] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.560898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.560928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.561175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.561203] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.561581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.561609] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.561972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.562007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.562362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.562390] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.562752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.562783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.563150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.563177] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.563548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.563576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.563952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.563981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.564344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.564371] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.564814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.564843] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.565197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.565226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.565481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.565510] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.565748] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.565781] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.566128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.566156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.566515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.566544] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.566916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.566947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.567304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.567332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.567721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.567752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.568130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.568158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.568521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.568549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.568920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.568949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.569317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.569345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.569590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.569619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.569976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.570006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.570369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.570397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.570765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.570794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.571011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.571043] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.571290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.571318] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.571675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.571712] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.572108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.572137] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.572427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.572454] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.572704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.572737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.925 [2024-10-17 17:50:05.573056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.925 [2024-10-17 17:50:05.573083] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.925 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.573438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.573466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.573829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.573860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.574232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.574260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.574527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.574555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.574923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.574953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.575371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.575400] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.575753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.575783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.576152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.576180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.576536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.576564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.576917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.576953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.577287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.577315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.577679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.577719] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.577975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.578008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.578365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.578394] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.578757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.578789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.579153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.579181] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.579544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.579573] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.579942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.579971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.580331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.580359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.580719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.580748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.581112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.581140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.581551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.581580] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.581946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.581983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.582333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.582362] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.582711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.582743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.582971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.583003] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.583437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.583466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.583835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.583865] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.584236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.584264] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.584523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.584551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.584816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.584848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.585133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.585162] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.585520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.585547] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.585924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.585952] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.926 qpair failed and we were unable to recover it. 00:28:57.926 [2024-10-17 17:50:05.586307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.926 [2024-10-17 17:50:05.586335] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.586732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.586764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.587118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.587147] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.587519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.587547] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.587869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.587898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.588263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.588291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.588631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.588658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.589035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.589064] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.589430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.589458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.589821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.589850] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.590258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.590285] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.590646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.590674] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.591134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.591163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.591516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.591544] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.591954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.591984] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.592343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.592378] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.592741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.592770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.592928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.592960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.593342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.593370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.593740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.593770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.594167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.594194] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.594413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.594444] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.594805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.594835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.595207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.595236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.595606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.595633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.596006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.596036] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.596393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.596421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.596794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.596823] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.597207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.597235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.597595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.597624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.597989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.598018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.927 [2024-10-17 17:50:05.598381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.927 [2024-10-17 17:50:05.598410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.927 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.598777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.598806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.599165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.599193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.599559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.599587] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.599956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.599985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.600347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.600377] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.600752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.600782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.601042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.601070] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.601427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.601454] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.601823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.601853] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.602109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.602140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.602528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.602557] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.602917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.602946] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.603308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.603336] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.603747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.603776] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.604124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.604154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.604531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.604559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.604836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.604866] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.605216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.605244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.605608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.605636] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.605993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.606023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.606379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.606408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.606663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.606711] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.607067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.607096] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.607452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.607488] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.607855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.607884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.608124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.608152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.608520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.608548] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.608917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.608947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.609308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.609335] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.609706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.609735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.610142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.610171] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.610537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.610565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.610998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.611027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.611380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.928 [2024-10-17 17:50:05.611408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.928 qpair failed and we were unable to recover it. 00:28:57.928 [2024-10-17 17:50:05.611771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.611801] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.612163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.612191] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.612551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.612578] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.612919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.612951] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.613303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.613331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.613704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.613733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.614139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.614168] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.614532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.614560] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.614926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.614955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.615334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.615362] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.615726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.615755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.616007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.616035] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.616423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.616451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.616702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.616731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.617136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.617164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.617526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.617554] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.617944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.617973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.618339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.618367] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.618735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.618766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.619019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.619047] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.619391] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.619426] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.619783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.619812] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.620171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.620200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.620567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.620595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.620970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.621000] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.621356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.621384] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.621753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.621784] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.622188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.622216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.622581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.622609] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.622847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.622900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.623285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.623314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.623726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.623755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.624171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.624198] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.929 qpair failed and we were unable to recover it. 00:28:57.929 [2024-10-17 17:50:05.624529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.929 [2024-10-17 17:50:05.624559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.624928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.624958] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.625323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.625351] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.625715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.625745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.626123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.626150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.626464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.626491] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.626874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.626903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.627259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.627289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.627589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.627616] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.627959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.627988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.628253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.628281] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.628668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.628706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.629093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.629121] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.629488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.629516] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.629876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.629905] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.630281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.630309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.630681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.630731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.631091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.631120] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.631483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.631512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.631876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.631905] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.632230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.632258] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.632630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.632657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.633109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.633138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.633510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.633538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.633784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.633813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.634056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.634084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.634461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.634488] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.634836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.634866] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.635233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.635260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.635641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.635668] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.636077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.636106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.636335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.636363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.636593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.636623] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.637012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.637043] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.930 [2024-10-17 17:50:05.637437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.930 [2024-10-17 17:50:05.637465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.930 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.637834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.637864] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.638221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.638257] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.638623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.638650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.638948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.638977] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.639327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.639355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.639713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.639743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.640111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.640140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.640515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.640543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.640892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.640921] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.641327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.641354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.641714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.641745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.642030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.642058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.642408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.642437] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.642819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.642848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.643208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.643236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.643457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.643487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.643751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.643782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.644085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.644115] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.644480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.644509] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.644855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.644886] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.645256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.645284] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.645602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.645631] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.646051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.646082] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.646421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.646449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.646816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.646846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.647215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.647244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.647478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.647506] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.647853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.647881] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.648274] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.648304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.648655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.648684] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.649054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.649084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.649449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.649477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.649843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.931 [2024-10-17 17:50:05.649872] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.931 qpair failed and we were unable to recover it. 00:28:57.931 [2024-10-17 17:50:05.650253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.650281] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.650535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.650563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.650837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.650869] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.651311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.651340] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.651707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.651737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.652094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.652123] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.652492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.652521] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.652951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.652983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.653238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.653274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.653641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.653669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.654044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.654073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.654435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.654467] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.654689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.654742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.655187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.655216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.655578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.655606] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.655948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.655979] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.656343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.656374] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.656740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.656770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.657145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.657174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.657576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.657605] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.657936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.657966] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.658334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.658363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.658728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.658758] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.659201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.659229] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.659567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.659595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.660021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.660052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.660402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.660430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.660816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.660846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.661223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.661251] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.661623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.661652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.662044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.662074] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.662417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.662445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.932 qpair failed and we were unable to recover it. 00:28:57.932 [2024-10-17 17:50:05.662758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.932 [2024-10-17 17:50:05.662787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.663166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.663194] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.663557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.663588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.663985] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.664018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.664379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.664409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.664664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.664704] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.665096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.665126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.665494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.665524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.665871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.665900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.666267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.666297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.666660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.666688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.667060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.667090] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.667494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.667523] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.667753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.667782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.668135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.668164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.668614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.668643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.668939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.668976] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.669313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.669342] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.669683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.669724] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.670076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.670104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.670472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.670502] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.670860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.670890] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.671257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.671286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.671646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.671675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.672046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.672075] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.672439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.672469] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.672831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.672862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.673220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.673248] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.673618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.673647] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.674010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.674042] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.674397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.674427] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.674786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.674817] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.675203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.675231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.675588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.675617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.933 [2024-10-17 17:50:05.675977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.933 [2024-10-17 17:50:05.676008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.933 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.676367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.676396] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.676766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.676798] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.677162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.677193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.677563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.677592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.677970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.678001] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.678363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.678393] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.678738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.678768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.679137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.679166] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.679525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.679555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.679894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.679923] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.680280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.680309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.680489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.680517] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.680889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.680919] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.681280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.681308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.681664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.681746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.682080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.682115] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.682365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.682394] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.682779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.682810] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.683214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.683242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.683485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.683517] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.683831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.683862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.684253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.684290] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.684613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.684642] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.684917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.684948] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.685188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.685219] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.685591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.685619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.685989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.686019] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.686381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.686410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.686772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.686802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.687164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.687193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.687561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.687591] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.687950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.687979] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.688352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.688380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.688814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.688844] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.689166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.689196] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.689546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.934 [2024-10-17 17:50:05.689576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.934 qpair failed and we were unable to recover it. 00:28:57.934 [2024-10-17 17:50:05.689838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.689868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.690235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.690263] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.690570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.690599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.690965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.690994] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.691419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.691447] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.691704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.691735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.692133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.692161] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.692522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.692550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.692907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.692936] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.693296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.693325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.693684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.693727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.694089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.694118] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.694479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.694510] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.694873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.694904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.695265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.695293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.695653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.695682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.695928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.695961] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.696337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.696365] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.696736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.696767] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.697146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.697176] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.697537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.697567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.697837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.697867] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.698253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.698282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.698715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.698746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.699103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.699131] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.699468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.699502] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.699841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.699872] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.700226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.700254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.700672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.700714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.701070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.701099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.701533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.701561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.701924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.701953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.702323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.702351] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.702730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.702762] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.935 qpair failed and we were unable to recover it. 00:28:57.935 [2024-10-17 17:50:05.703119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.935 [2024-10-17 17:50:05.703149] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.703508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.703537] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.703883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.703912] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.704279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.704308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.704663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.704703] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.705094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.705125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.705492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.705524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.705870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.705900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.706212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.706242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.706608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.706636] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.707017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.707049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.707407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.707436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.707817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.707849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.708208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.708238] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.708601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.708632] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.708991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.709022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.709294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.709322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.709706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.709737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.710114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.710145] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.710575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.710603] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.710944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.710973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.711325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.711353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.711721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.711752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.712144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.712172] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.712554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.712583] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.712812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.712840] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.713195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.713224] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.713590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.713619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.714011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.714042] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.714475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.714511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.714862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.714893] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.715266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.936 [2024-10-17 17:50:05.715302] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.936 qpair failed and we were unable to recover it. 00:28:57.936 [2024-10-17 17:50:05.715677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.715720] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.716074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.716102] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.716462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.716490] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.716762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.716791] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.717197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.717225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.717585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.717613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.717959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.717988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.718355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.718383] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.718749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.718778] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.719153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.719181] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.719539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.719567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.719940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.719969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.720385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.720412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.720645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.720673] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.721069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.721099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.721318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.721345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.721726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.721757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.722153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.722183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.722544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.722573] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.722927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.722957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.723325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.723353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.723569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.723598] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.723865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.723895] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.724238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.724266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.724511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.724538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.724881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.724910] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.725264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.725294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.725545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.725576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.725810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.725844] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.726212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.937 [2024-10-17 17:50:05.726240] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.937 qpair failed and we were unable to recover it. 00:28:57.937 [2024-10-17 17:50:05.726604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.726633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.726882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.726915] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.727271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.727301] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.727675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.727717] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.728068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.728097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.728463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.728491] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.728865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.728895] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.729251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.729279] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.729517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.729549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.729935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.729971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.730308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.730336] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.730563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.730593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.730829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.730863] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.731231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.731260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.731508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.731537] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.731877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.731906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.732270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.732299] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.732555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.732583] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.732956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.732985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.733360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.733388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.733749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.733780] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.734153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.734181] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.734549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.734577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.734947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.734977] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.735318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.735346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.735716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.735745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.736080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.736108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.736409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.736436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.736683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.736727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.737100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.737129] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.737485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.737512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.737875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.737906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.738304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.738332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.738704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.738733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.739136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.938 [2024-10-17 17:50:05.739164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.938 qpair failed and we were unable to recover it. 00:28:57.938 [2024-10-17 17:50:05.739511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.739540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.739913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.739942] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.740305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.740334] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.740579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.740611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.740958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.740986] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.741411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.741439] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.741783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.741814] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.742176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.742204] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.742561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.742588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.742950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.742979] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.743350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.743377] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.743747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.743777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.744145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.744173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.744620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.744647] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.744986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.745016] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.745377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.745406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.745768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.745799] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.746141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.746169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.746530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.746557] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.746935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.746963] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.747340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.747369] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.747731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.747760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.748146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.748174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.748544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.748574] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.748794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.748826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.749190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.749218] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.749625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.749654] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.750030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.750059] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.750309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.750340] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.750714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.750743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.751152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.751181] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.751533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.751562] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.751918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.751948] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.752306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.752334] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.752704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.752734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.939 qpair failed and we were unable to recover it. 00:28:57.939 [2024-10-17 17:50:05.753096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.939 [2024-10-17 17:50:05.753124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.753488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.753518] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.753764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.753794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.754156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.754184] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.754567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.754595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.754808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.754839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.755222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.755257] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.755510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.755539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.755903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.755933] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.756291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.756320] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.756622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.756653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.757022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.757052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.757413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.757442] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.757821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.757851] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.758227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.758255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.758664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.758704] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.759062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.759091] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.759450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.759479] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.759869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.759899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.760253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.760282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.760529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.760560] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.760809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.760838] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.761297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.761325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.761656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.761685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.762051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.762079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.762454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.762482] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.762854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.762884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.763215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.763244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.763506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.763535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.763870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.763900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.764264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.764293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.764657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.764685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.765024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.940 [2024-10-17 17:50:05.765053] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.940 qpair failed and we were unable to recover it. 00:28:57.940 [2024-10-17 17:50:05.765470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.765499] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.765739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.765773] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.766022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.766049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.766294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.766324] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.766711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.766742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.767084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.767113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.767487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.767517] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.767880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.767910] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.768277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.768305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.768665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.768718] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.769084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.769112] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.769470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.769500] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.769849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.769878] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.770238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.770272] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.770639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.770667] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.771052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.771081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.771445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.771473] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.771720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.771750] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.772111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.772139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.772509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.772536] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.772788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.772820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.773201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.773230] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.773594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.773623] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.774052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.774082] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.774442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.774471] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.774828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.774857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.775239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.775267] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.775635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.775664] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.776027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.776057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.776417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.776445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.776714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.776744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.777106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.777134] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.777499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.777526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.777874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.777904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.778343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.778372] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.778740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.941 [2024-10-17 17:50:05.778770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.941 qpair failed and we were unable to recover it. 00:28:57.941 [2024-10-17 17:50:05.779142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.779170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.779431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.779458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.779814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.779843] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.780194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.780223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.780587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.780616] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.780957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.780988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.781222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.781253] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.781487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.781517] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.781729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.781760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.782149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.782178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.782546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.782575] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.782930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.782960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.783326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.783355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.783722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.783752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.784116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.784144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.784510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.784538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.784917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.784947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.785323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.785357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.785781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.785810] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.786172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.786200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.786545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.786573] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.786927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.786956] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.787220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.787248] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.787486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.787516] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.787919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.787948] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.788310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.788338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.788677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.788717] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.789080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.789108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.789479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.789507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.789872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.789903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.942 [2024-10-17 17:50:05.790336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.942 [2024-10-17 17:50:05.790364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.942 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.790590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.790620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.790996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.791025] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.791396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.791425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.791784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.791813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.792168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.792196] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.792572] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.792599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.792970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.793000] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.793363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.793390] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.793728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.793757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.794137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.794166] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.794527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.794554] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.794802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.794834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.795192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.795220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.795591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.795619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.795986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.796015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.796359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.796387] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.796747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.796775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.797151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.797178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.797523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.797551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.797818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.797848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.798209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.798236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.798606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.798633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.798994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.799023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.799403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.799431] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.799783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.799812] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.800175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.800203] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.800542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.800575] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.800936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.800967] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.801336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.801364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.801713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.801742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.801950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.801980] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.802266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.943 [2024-10-17 17:50:05.802294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.943 qpair failed and we were unable to recover it. 00:28:57.943 [2024-10-17 17:50:05.802647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.802676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.803059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.803088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.803451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.803478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.803747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.803777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.804130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.804160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.804509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.804537] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.804907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.804937] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.805353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.805382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.805721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.805753] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.806117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.806146] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.806504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.806533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.806920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.806950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.807317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.807347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.807720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.807751] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.808119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.808147] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.808509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.808538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.808909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.808937] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.809283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.809311] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.809665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.809702] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.810061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.810089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.810437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.810465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.810834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.810864] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.811227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.811254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.811622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.811650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.812022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.812052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.812431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.812458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.812819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.812850] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.813218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.813247] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.813651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.813679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.814058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.814089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.814446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.814475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.814850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.814880] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.815125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.815155] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.815519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.815548] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.815891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.815932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.816290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.816318] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.816686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.816728] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.817115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.817143] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.817500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.944 [2024-10-17 17:50:05.817527] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.944 qpair failed and we were unable to recover it. 00:28:57.944 [2024-10-17 17:50:05.817905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.817935] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.818344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.818372] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.818735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.818764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.819117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.819145] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.819504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.819533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.819882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.819910] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.820306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.820333] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.820720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.820750] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.821115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.821144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.821507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.821535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.821920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.821950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.822327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.822356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.822731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.822761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.823122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.823151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.823552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.823580] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.823929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.823958] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.824212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.824240] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.824642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.824670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.825027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.825056] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.825417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.825446] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.825818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.825848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.826204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.826233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.826558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.826587] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.826926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.826956] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.827298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.827326] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.827712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.827741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:57.945 [2024-10-17 17:50:05.828096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:57.945 [2024-10-17 17:50:05.828123] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:57.945 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.828479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.828511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.828743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.828775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.829183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.829211] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.829570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.829598] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.829956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.829986] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.830358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.830386] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.830763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.830792] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.831166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.831195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.831584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.831619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.831860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.831888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.832236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.832264] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.832590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.832619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.833028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.833057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.833422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.833451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.833814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.833844] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.834218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.834246] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.834699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.834730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.834974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.835004] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.835375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.835403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.835746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.835784] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.836160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.836187] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.836549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.836577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.836944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.836974] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.837324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.837352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.837720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.837750] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.838098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.838126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.838511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.838538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.838905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.838935] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.839305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.839333] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.839703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.839732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.840105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.840135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.840506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.840534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.840890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.840920] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.841294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.841322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.235 [2024-10-17 17:50:05.841713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.235 [2024-10-17 17:50:05.841743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.235 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.842149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.842178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.842414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.842446] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.842743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.842773] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.843129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.843158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.843506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.843534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.843891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.843921] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.844165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.844197] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.844552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.844581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.844931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.844960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.845319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.845347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.845723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.845752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.846109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.846137] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.846506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.846534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.846875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.846911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.847063] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.847092] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.847505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.847533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.847883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.847913] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.848271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.848299] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.848661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.848689] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.849077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.849106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.849360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.849389] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.849738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.849768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.850146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.850174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.850424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.850452] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.850805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.850836] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.851210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.851237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.851602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.851630] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.851875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.851908] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.852282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.852310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.852673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.852710] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.853143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.853172] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.853567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.853595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.853931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.853961] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.854327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.854355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.854721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.854751] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.855114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.855142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.855496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.855525] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.855875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.855904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.856305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.856333] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.856680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.856726] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.857091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.236 [2024-10-17 17:50:05.857120] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.236 qpair failed and we were unable to recover it. 00:28:58.236 [2024-10-17 17:50:05.857483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.857511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.857871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.857899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.858264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.858292] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.858655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.858684] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.858932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.858963] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.859163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.859190] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.859586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.859615] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.859975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.860006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.860364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.860393] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.860795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.860824] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.861163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.861191] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.861568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.861596] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.862013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.862048] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.862390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.862419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.862797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.862826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.863188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.863215] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.863657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.863685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.863981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.864009] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.864376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.864405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.864772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.864803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.865236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.865263] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.865515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.865546] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.865906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.865936] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.866304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.866332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.866690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.866729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.867070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.867099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.867472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.867501] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.867864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.867894] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.868060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.868092] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.868459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.868486] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.868849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.868880] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.237 [2024-10-17 17:50:05.869230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.237 [2024-10-17 17:50:05.869258] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.237 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.869521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.869549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.869913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.869942] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.870324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.870353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.870617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.870645] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.870925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.870955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.871368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.871397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.871768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.871797] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.872064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.872096] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.872385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.872412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.872771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.872802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.873142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.873170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.873543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.873572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.873972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.874002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.874366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.874394] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.874723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.874752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.875114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.875142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.875497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.875526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.875900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.875929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.876289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.876318] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.876684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.876736] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.877086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.877121] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.877532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.877560] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.877938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.877969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.878317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.878345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.878708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.878739] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.879092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.879120] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.879419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.879446] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.879668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.879709] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.880143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.880171] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.880526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.880555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.880923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.880953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.881313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.881341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.881713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.881742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.238 [2024-10-17 17:50:05.882114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.238 [2024-10-17 17:50:05.882142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.238 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.882513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.882541] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.882929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.882958] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.883331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.883359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.883654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.883682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.884054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.884083] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.884447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.884474] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.884840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.884869] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.885240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.885268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.885629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.885656] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.886030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.886061] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.886406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.886434] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.886807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.886837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.887177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.887205] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.887570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.887602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.887989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.888018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.888385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.888415] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.888788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.888818] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.889191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.889220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.889581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.889612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.889789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.889821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.890199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.890227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.890596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.890625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.891007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.891038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.891369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.891398] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.891759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.891788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.892159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.892187] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.239 [2024-10-17 17:50:05.892553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.239 [2024-10-17 17:50:05.892594] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.239 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.892953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.892984] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.893342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.893371] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.893722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.893755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.894117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.894146] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.894585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.894612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.894984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.895016] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.895279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.895308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.895652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.895680] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.896061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.896092] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.896452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.896481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.896845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.896875] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.897242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.897271] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.897686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.897734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.898006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.898039] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.898431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.898461] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.898902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.898933] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.899299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.899329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.899700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.899731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.900087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.900116] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.900471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.900499] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.900844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.900874] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.901136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.901165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.901394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.901425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.901772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.901803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.902171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.902200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.902544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.902575] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.902812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.902843] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.903245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.903274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.903490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.903518] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.903778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.903807] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.904226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.904255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.904589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.904617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.904989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.905018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.905383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.905413] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.905800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.905830] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.906193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.906222] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.906608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.906637] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.906977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.907010] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.907359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.907388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.907750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.907788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.908125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.908154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.908523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.908551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.908915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.908945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.909315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.909345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.909713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.909742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.910113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.910142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.910371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.910403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.910780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.910811] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.911094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.911123] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.911461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.911490] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.911848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.911878] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.912122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.912153] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.912490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.912518] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.912853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.912891] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.913291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.913322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.913723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.913755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.914116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.914145] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.914507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.914536] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.914903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.914933] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.915289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.915318] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.915677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.915720] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.240 qpair failed and we were unable to recover it. 00:28:58.240 [2024-10-17 17:50:05.916077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.240 [2024-10-17 17:50:05.916105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.916466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.916495] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.916864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.916895] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.917265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.917294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.917660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.917689] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.918106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.918136] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.918499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.918528] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.918883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.918914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.919277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.919306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.919543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.919576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.919951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.919983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.920330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.920358] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.920735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.920766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.921158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.921187] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.921551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.921581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.921960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.921990] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.922358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.922387] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.922753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.922782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.923143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.923179] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.923524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.923552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.923926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.923957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.924300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.924330] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.924762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.924794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.925145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.925175] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.925540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.925569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.925910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.925940] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.926298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.926329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.926689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.926732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.927094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.927124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.927456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.927485] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.927855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.927886] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.928231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.928259] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.928624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.928653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.929020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.929051] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.929407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.929436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.929801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.929832] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.930179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.930208] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.930463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.930492] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.930874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.930904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.931265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.931294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.931658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.931686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.932051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.932081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.932454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.932483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.932852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.932884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.933249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.933278] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.933648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.933679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.934077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.934108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.934334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.934367] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.934740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.934773] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.935149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.935181] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.935508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.935537] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.935869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.935900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.936160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.936193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.936547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.936577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.936848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.936879] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.937230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.937260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.937626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.937655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.938047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.938078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.938434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.938464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.938831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.938862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.939116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.939146] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.939500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.939530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.939864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.241 [2024-10-17 17:50:05.939895] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.241 qpair failed and we were unable to recover it. 00:28:58.241 [2024-10-17 17:50:05.940261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.940291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.940663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.940703] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.941064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.941092] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.941326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.941356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.941718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.941748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.942154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.942183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.942539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.942568] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.942941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.942972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.943335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.943364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.943727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.943758] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.944118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.944149] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.944511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.944540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.944917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.944947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.945302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.945331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.945687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.945732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.946078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.946107] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.946469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.946498] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.946868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.946900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.947259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.947288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.947651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.947681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.948074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.948104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.948468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.948496] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.948862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.948897] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.949255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.949285] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.949651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.949681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.950046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.950075] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.950438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.950466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.950832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.950863] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.951244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.951272] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.951604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.951634] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.952031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.952061] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.952495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.952523] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.952892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.952922] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.953282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.953309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.953672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.953711] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.954049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.954079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.954446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.954475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.954719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.954763] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.955121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.955150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.955522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.955549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.955956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.955987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.956221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.956253] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.956607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.956636] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.956981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.957011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.957278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.957306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.957727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.957757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.958124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.958152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.958592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.958621] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.958977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.959007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.242 [2024-10-17 17:50:05.959368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.242 [2024-10-17 17:50:05.959397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.242 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.959636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.959663] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.959926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.959955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.960318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.960346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.960717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.960748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.961108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.961136] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.961454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.961482] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.961871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.961900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.962261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.962291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.962651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.962679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.963051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.963080] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.963522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.963550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.963925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.963954] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.964294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.964327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.964684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.964727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.965070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.965107] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.965479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.965507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.965857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.965887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.966262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.966291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.966658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.966688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.967046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.967075] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.967313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.967344] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.967725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.967756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.968113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.968143] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.968520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.968549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.968914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.968943] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.969314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.969342] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.969723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.969753] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.970120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.970150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.970390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.970418] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.970723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.970752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.971119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.971148] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.971521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.971549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.971921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.971951] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.972329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.972357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.972728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.972760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.973131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.973161] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.973525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.973555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.973920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.973950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.974315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.974343] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.974714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.974744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.975106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.975134] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.975502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.975530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.975787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.975816] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.976192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.976221] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.976647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.976675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.976965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.976994] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.977345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.977374] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.977823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.977852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.978106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.978133] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.978347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.978375] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.978745] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.978775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.979138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.979167] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.979533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.979570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.979950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.979979] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.980357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.980386] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.980750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.980780] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.981143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.981171] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.981533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.981562] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.981943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.243 [2024-10-17 17:50:05.981972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.243 qpair failed and we were unable to recover it. 00:28:58.243 [2024-10-17 17:50:05.982333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.982360] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.982727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.982757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.983106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.983135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.983507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.983535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.983897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.983927] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.984313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.984341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.984737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.984766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.985129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.985158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.985501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.985529] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.985896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.985927] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.986169] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.986198] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.986558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.986588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.986952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.986981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.987343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.987371] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.987732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.987762] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.987895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.987927] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.988260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.988289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.988542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.988574] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.988837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.988867] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.989304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.989333] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.989709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.989741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.990093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.990121] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.990554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.990582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.990917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.990948] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.991313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.991341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.991709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.991739] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.992078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.992107] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.992356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.992387] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.992738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.992769] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.993188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.993216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.993504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.993532] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.993950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.993979] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.994416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.994445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.994814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.994851] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.995229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.995257] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.995641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.995669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.995908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.995941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.996282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.996311] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.996603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.996632] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.996986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.997016] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.997385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.997413] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.997721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.997751] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.998123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.998151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.998513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.998542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.998917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.998947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.999304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.999333] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.999552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.999583] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:05.999939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:05.999968] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.000308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.000336] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.000705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.000735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.001085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.001113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.001483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.001512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.001884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.001915] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.002274] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.002302] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.002666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.002705] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.002948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.002978] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.003356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.003385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.003769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.003800] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.004168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.004197] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.004552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.004581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.004956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.004988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.005255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.005282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.005630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.244 [2024-10-17 17:50:06.005657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.244 qpair failed and we were unable to recover it. 00:28:58.244 [2024-10-17 17:50:06.006027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.006058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.006428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.006455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.006835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.006866] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.007221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.007250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.007633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.007662] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.008032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.008062] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.008418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.008449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.008818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.008848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.009221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.009250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.009600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.009627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.009877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.009916] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.010277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.010306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.010678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.010718] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.011078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.011107] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.011474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.011502] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.011876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.011906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.012264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.012293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.012649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.012678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.013053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.013082] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.013440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.013468] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.013838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.013868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.014112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.014143] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.014489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.014516] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.014905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.014935] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.015173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.015204] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.015561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.015590] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.016012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.016042] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.016304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.016332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.016730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.016761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.017159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.017188] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.017435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.017464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.017821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.017852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.018211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.018243] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.018607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.018636] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.018943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.018972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.019205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.019237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.019611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.019649] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.020048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.020081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.020421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.020450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.020798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.020831] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.021192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.021220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.021585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.021613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.021980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.022011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.022379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.022407] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.022621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.022652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.023054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.023084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.023446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.023475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.023849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.023879] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.024141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.024169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.024524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.024553] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.024915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.024952] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.025379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.025408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.025653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.025681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.026088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.026119] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.026286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.026317] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.026549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.026582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.245 qpair failed and we were unable to recover it. 00:28:58.245 [2024-10-17 17:50:06.026934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.245 [2024-10-17 17:50:06.026964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.027348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.027376] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.027736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.027765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.028031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.028061] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.028400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.028428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.028842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.028873] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.029219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.029249] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.029583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.029611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.029962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.029992] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.030358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.030386] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.030751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.030783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.031172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.031200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.031566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.031595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.031952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.031982] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.032339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.032368] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.032737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.032768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.032989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.033021] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.033374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.033403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.033784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.033814] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.034229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.034257] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.034617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.034644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.035045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.035074] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.035433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.035461] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.035827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.035856] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.036226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.036254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.036626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.036653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.036937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.036969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.037341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.037369] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.037729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.037760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.038113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.038141] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.038381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.038411] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.038653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.038685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.039068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.039099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.039314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.039347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.039750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.039786] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.040145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.040176] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.040523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.040555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.040955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.040985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.041341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.041371] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.041730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.041764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.042138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.042168] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.042529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.042558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.042935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.042965] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.043170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.043202] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.043542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.043570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.043943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.043975] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.044399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.044427] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.044776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.044806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.045250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.045280] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.045533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.045565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.045808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.045838] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.046218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.046245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.046608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.046637] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.046869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.046899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.047285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.047314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.047684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.047727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.048101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.048129] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.048490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.048519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.048876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.048906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.049175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.049203] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.049583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.049612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.246 [2024-10-17 17:50:06.049850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.246 [2024-10-17 17:50:06.049884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.246 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.050258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.050288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.050648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.050680] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.051045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.051075] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.051327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.051357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.051727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.051757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.052106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.052134] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.052482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.052512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.052855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.052887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.053229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.053257] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.053665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.053706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.054134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.054165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.054402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.054431] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.054762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.054799] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.055157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.055186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.055559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.055587] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.055930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.055960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.056304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.056332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.056596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.056626] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.056982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.057013] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.057380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.057410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.057778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.057808] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.058178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.058206] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.058563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.058591] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.058986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.059015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.059372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.059412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.059671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.059713] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.060069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.060099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.060451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.060481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.060838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.060868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.061225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.061255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.061615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.061643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.061990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.062020] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.062375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.062405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.062768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.062798] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.063227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.063256] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.063508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.063536] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.063870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.063900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.064277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.064308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.064716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.064748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.065074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.065104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.065482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.065510] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.065892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.065923] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.066199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.066230] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.066585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.066615] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.066987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.067016] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.067383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.067413] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.067680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.067721] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.068100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.068130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.068437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.068466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.068612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.068651] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.069049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.069079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.069427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.069457] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.069826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.069862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.070200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.070229] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.070491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.070522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.070796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.070826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.071205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.071235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.071515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.071543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.071892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.071922] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.072300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.247 [2024-10-17 17:50:06.072329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.247 qpair failed and we were unable to recover it. 00:28:58.247 [2024-10-17 17:50:06.072756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.072787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.073166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.073195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.073539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.073567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.073840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.073870] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.074234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.074264] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.074627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.074655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.074946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.074976] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.075334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.075363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.075719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.075749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.076142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.076170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.076294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.076323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.076704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.076734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.076975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.077006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.077253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.077282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.077640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.077671] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.078033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.078063] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.078409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.078445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.078831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.078861] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.079232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.079261] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.079614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.079642] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.079924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.079958] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.080314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.080343] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.080581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.080622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.081014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.081044] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.081407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.081436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.081802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.081832] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.082179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.082208] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.082565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.082594] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.082935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.082972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.083304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.083332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.083667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.083706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.084045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.084073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.084390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.084426] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.084772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.084802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.085169] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.085197] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.085505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.085533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.085903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.085932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.086189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.086216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.086557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.086586] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.086953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.086983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.087281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.087309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.087668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.087716] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.088075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.088104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.088473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.088503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.088897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.088927] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.089276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.089306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.089759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.089789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.090131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.090159] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.090449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.090477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.090855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.090885] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.091111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.091139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.091528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.091556] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.248 [2024-10-17 17:50:06.091950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.248 [2024-10-17 17:50:06.091981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.248 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.092327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.092355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.092654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.092682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.093065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.093094] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.093337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.093365] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.093746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.093774] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.094155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.094183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.094458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.094487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.094935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.094965] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.095208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.095236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.095585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.095614] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.096027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.096057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.096411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.096441] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.096776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.096804] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.097086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.097114] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.097258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.097289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.097616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.097644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.097911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.097941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.098309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.098337] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.098707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.098738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.099099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.099134] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.099499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.099528] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.099875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.099904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.100292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.100321] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.100668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.100707] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.101077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.101105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.101448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.101477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.101910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.101938] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.102301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.102329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.102712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.102742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.103127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.103155] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.103385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.103415] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.103644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.103672] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.104073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.104103] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.104490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.104519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.104915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.104945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.105321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.105348] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.105726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.105755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.106124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.106151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.106507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.106534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.106745] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.106774] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.107035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.107062] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.107422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.107451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.107796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.107825] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.108083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.108113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.108447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.108475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.108727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.108756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.109136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.109164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.109529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.109558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.109999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.110028] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.110240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.110268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.110659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.110687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.111060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.111088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.111441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.111469] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.111817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.111847] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.112225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.112253] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.112614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.112642] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.113059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.113089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.113485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.249 [2024-10-17 17:50:06.113513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.249 qpair failed and we were unable to recover it. 00:28:58.249 [2024-10-17 17:50:06.113888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.113919] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.114289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.114323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.114550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.114581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.114933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.114962] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.115328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.115357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.115726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.115757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.116128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.116156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.116530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.116558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.117009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.117038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.117394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.117421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.117760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.117791] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.118176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.118204] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.118578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.118606] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.118958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.118987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.119351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.119379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.119775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.119804] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.120153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.120183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.120560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.120589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.120934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.120962] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.121341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.121368] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.121740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.121770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.122025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.122053] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.122303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.122331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.122707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.122738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.124689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.124767] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.125096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.125127] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.125474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.125503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.125849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.125881] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.126258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.126288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.250 [2024-10-17 17:50:06.126650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.250 [2024-10-17 17:50:06.126678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.250 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.127067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.127097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.127447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.127477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.127820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.127850] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.128202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.128231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.128600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.128628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.129080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.129109] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.129555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.129583] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.129953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.129991] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.130284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.130311] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.130700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.130730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.130998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.131030] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.131383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.131419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.131789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.131818] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.132188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.132217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.132475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.132504] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.132873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.132902] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.133257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.133286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.522 [2024-10-17 17:50:06.133734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.522 [2024-10-17 17:50:06.133765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.522 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.134133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.134161] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.134510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.134538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.134925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.134955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.135218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.135246] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.135626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.135653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.136111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.136140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.136494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.136522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.136886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.136916] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.137279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.137307] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.137670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.137707] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.137980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.138008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.138349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.138379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.138772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.138802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.139207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.139235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.139568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.139595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.139872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.139903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.140220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.140251] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.140644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.140673] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.140948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.140977] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.141260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.141288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.141640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.141675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.142150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.142183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.142323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.142354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.142610] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.142638] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.143038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.143070] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.143413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.143440] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.143727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.143758] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.144136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.144166] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.144513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.144542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.145000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.145030] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.145379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.145408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.145700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.145730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.146080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.146109] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.146331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.146364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.146751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.146783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.147042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.147072] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.147405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.147435] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.147685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.147729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.148120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.148150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.148499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.148528] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.523 qpair failed and we were unable to recover it. 00:28:58.523 [2024-10-17 17:50:06.148913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.523 [2024-10-17 17:50:06.148943] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.149185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.149215] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.149549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.149578] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.149951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.149983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.150362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.150391] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.150680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.150723] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.150980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.151011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.151397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.151427] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.151814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.151845] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.152219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.152250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.152602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.152633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.152908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.152940] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.153288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.153317] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.153688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.153732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.154162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.154191] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.154540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.154569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.154755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.154788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.155174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.155204] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.155586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.155616] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.155957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.155987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.156392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.156429] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.156780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.156811] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.157061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.157095] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.157464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.157494] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.157758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.157788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.158069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.158099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.158426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.158455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.158723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.158757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.159152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.159184] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.159522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.159552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.159921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.159952] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.160233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.160263] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.160505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.160534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.160651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.160680] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.161074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.161106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.161452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.161483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.161870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.161900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.162278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.162309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.162565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.162595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.162866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.524 [2024-10-17 17:50:06.162898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.524 qpair failed and we were unable to recover it. 00:28:58.524 [2024-10-17 17:50:06.163274] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.163303] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.163673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.163715] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.163972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.164001] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.164383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.164412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.164711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.164742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.165134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.165163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.165417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.165450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.165684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.165727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.166081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.166110] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.166368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.166397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.166750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.166782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.167225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.167253] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.167633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.167663] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.168049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.168079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.168434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.168464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.168844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.168876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.169137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.169169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.169389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.169420] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.169739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.169770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.170119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.170148] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.170496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.170530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.170881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.170914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.171266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.171295] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.173456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.173526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.173956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.173994] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.174249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.174282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.174565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.174595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.174861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.174892] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.175276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.175304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.175669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.175730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.176128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.176158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.176528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.176557] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.176884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.176914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.177205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.177233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.177602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.177631] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.178009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.178040] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.178404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.178433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.178797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.178827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.179204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.179232] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.179536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.525 [2024-10-17 17:50:06.179564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.525 qpair failed and we were unable to recover it. 00:28:58.525 [2024-10-17 17:50:06.179934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.179964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.180357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.180387] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.180756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.180786] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.181154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.181183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.181526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.181554] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.181917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.181946] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.182203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.182231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.182665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.182706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.183065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.183093] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.183318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.183351] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.183726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.183757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.184000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.184028] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.184389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.184418] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.184791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.184821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.185177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.185207] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.185569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.185597] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.189753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.189822] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.190336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.190379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.190569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.190604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.191019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.191059] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.191390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.191480] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.191906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.191946] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.192392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.192424] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.192810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.192842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.193215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.193245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.193619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.193648] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.193975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.194008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.194425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.194455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.194872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.194904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.195271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.195300] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.195678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.195718] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.195943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.195977] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.196389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.196419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.196808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.196841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.197222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.526 [2024-10-17 17:50:06.197252] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.526 qpair failed and we were unable to recover it. 00:28:58.526 [2024-10-17 17:50:06.197613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.197643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.197902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.197932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.198334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.198364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.198743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.198772] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.199187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.199217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.199472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.199501] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.199879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.199909] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.200299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.200318] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.202938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.202983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.203280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.203305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.203680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.203714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.204053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.204073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.204389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.204409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.204763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.204790] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.205149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.205169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.205522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.205543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.205903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.205924] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.206291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.206310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.206684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.206714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.207077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.207096] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.207472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.207492] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.207843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.207865] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.208109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.208129] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.208533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.208553] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.208882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.208903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.209243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.209267] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.209643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.209662] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.210045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.210065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.210286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.210310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.210652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.210672] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.211120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.211141] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.211486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.211506] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.211826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.211846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.212184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.212206] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.212540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.212563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.212920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.212947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.213293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.213316] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.213559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.213582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.213970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.213993] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.214235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.214260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.527 [2024-10-17 17:50:06.214659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.527 [2024-10-17 17:50:06.214683] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.527 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.215058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.215081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.215346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.215368] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.215701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.215725] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.216045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.216068] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.216397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.216421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.216568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.216594] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.216948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.216974] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.217187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.217210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.217564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.217588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.217954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.217981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.218358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.218382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.218749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.218774] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.219102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.219126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.219465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.219488] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.219759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.219783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.220040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.220063] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.220278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.220301] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.220517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.220542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.220796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.220820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.221214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.221237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.221481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.221505] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.221863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.221887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.222257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.222280] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.222665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.222687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.223056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.223085] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.223329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.223352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.223706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.223731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.224107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.224130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.224488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.224511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.224855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.224880] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.225229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.225252] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.225626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.225650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.226048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.226073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.226432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.226456] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.226815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.226839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.227205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.227233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.227600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.227628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.228027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.228058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.228412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.228440] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.228673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.528 [2024-10-17 17:50:06.228718] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.528 qpair failed and we were unable to recover it. 00:28:58.528 [2024-10-17 17:50:06.229089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.229117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.229482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.229518] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.229860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.229890] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.230258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.230286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.230655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.230683] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.231068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.231097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.231350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.231377] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.231687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.231727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.232114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.232142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.232502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.232530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.232916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.232945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.233214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.233242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.233627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.233655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.234258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.234289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.234515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.234543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.235135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.235165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.235516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.235544] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.235896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.235925] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.236295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.236323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.236708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.236737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.237100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.237130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.237465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.237494] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.237847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.237876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.238243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.238270] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.238518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.238552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.238813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.238844] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.239210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.239239] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.239601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.239631] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.239993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.240023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.240361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.240388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.240763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.240793] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.241145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.241181] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.241552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.241581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.241966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.241995] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.242352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.242380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.242736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.242766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.243037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.243069] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.243452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.243482] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.243728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.243762] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.244180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.244209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.529 [2024-10-17 17:50:06.244571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.529 [2024-10-17 17:50:06.244599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.529 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.244824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.244857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.245222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.245250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.245602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.245638] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.246041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.246073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.246413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.246451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.246705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.246735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.247110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.247139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.247388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.247419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.247781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.247811] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.248196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.248225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.248600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.248629] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.248997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.249027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.249389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.249416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.249772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.249801] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.250170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.250199] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.250449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.250481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.250947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.250978] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.251339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.251368] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.251735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.251766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.252022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.252053] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.252441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.252471] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.252855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.252886] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.253133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.253165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.253526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.253561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.253953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.253983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.254326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.254354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.254716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.254746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.254996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.255028] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.255258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.255289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.255549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.255579] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.255945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.255975] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.256406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.256434] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.256765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.256801] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.257173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.257201] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.257434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.257465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.257813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.257843] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.258224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.258252] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.258619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.258647] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.259021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.259051] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.259409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.259438] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.530 [2024-10-17 17:50:06.259808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.530 [2024-10-17 17:50:06.259838] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.530 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.260203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.260231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.260595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.260624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.260979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.261008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.261369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.261399] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.261781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.261811] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.262251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.262279] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.262641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.262670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.263033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.263062] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.263432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.263460] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.263836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.263867] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.264088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.264121] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.264485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.264514] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.264877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.264906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.265284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.265312] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.265672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.265724] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.266087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.266116] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.266481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.266509] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.266910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.266939] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.267294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.267322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.267686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.267728] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.268080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.268108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.268471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.268500] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.268872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.268910] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.269141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.269174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.269535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.269563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.269931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.269960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.270337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.270367] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.270579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.270612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.270967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.270996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.271245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.271277] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.271647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.271675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.272018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.272046] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.272297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.272328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.272705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.272735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.273085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.273115] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.531 [2024-10-17 17:50:06.273483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.531 [2024-10-17 17:50:06.273513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.531 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.273765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.273795] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.274159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.274187] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.274541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.274577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.274915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.274945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.275369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.275399] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.275571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.275602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.275969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.275998] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.276362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.276390] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.276760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.276789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.277152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.277180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.277542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.277571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.277941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.277971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.278330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.278359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.278733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.278764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.279132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.279161] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.279524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.279553] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.279914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.279950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.280356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.280383] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.280752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.280783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.281143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.281171] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.281540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.281570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.281940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.281971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.282311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.282349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.282709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.282740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.282993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.283025] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.283301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.283329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.283679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.283726] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.284129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.284158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.284392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.284424] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.284811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.284841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.285200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.285228] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.285579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.285606] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.285871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.285901] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.286258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.286286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.286649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.286678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.287044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.287073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.287438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.287465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.287832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.287861] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.288227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.288255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.288611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.532 [2024-10-17 17:50:06.288641] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.532 qpair failed and we were unable to recover it. 00:28:58.532 [2024-10-17 17:50:06.289003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.289034] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.289404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.289432] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.289681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.289737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.290138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.290166] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.290399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.290430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.290790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.290820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.291213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.291242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.291613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.291641] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.292004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.292034] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.292391] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.292420] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.292680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.292720] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.292976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.293009] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.293388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.293416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.293783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.293813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.294175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.294203] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.294560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.294587] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.294924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.294954] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.295318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.295348] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.295711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.295740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.296104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.296132] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.296497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.296525] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.296997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.297026] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.297386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.297415] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.297783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.297812] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.298180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.298208] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.298577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.298605] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.298972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.299007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.299351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.299388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.299765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.299795] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.300229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.300257] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.300616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.300645] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.300997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.301031] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.301294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.301321] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.301709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.301738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.301966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.301996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.302371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.302400] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.302769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.302800] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.303043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.303072] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.303429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.303456] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.303806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.533 [2024-10-17 17:50:06.303837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.533 qpair failed and we were unable to recover it. 00:28:58.533 [2024-10-17 17:50:06.304260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.304290] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.304653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.304682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.305078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.305106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.305280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.305310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.305668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.305707] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.306069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.306097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.306466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.306494] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.306742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.306772] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.307137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.307165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.307535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.307563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.307809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.307842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.308264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.308293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.308643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.308672] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.309083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.309114] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.309460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.309497] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.309792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.309821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.310182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.310210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.310571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.310598] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.310971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.311002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.311363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.311397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.311755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.311785] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.312193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.312221] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.312585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.312620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.312971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.313000] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.313351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.313380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.313633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.313665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.314082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.314118] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.314494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.314522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.314878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.314908] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.315308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.315338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.315684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.315723] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.316081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.316109] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.316422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.316450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.316802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.316832] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.317205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.317233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.317601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.317629] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.317992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.318023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.318362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.318390] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.318752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.318781] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.319053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.534 [2024-10-17 17:50:06.319084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.534 qpair failed and we were unable to recover it. 00:28:58.534 [2024-10-17 17:50:06.319480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.319511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.319771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.319800] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.320172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.320203] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.320573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.320602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.320991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.321020] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.321283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.321310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.321707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.321737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.322141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.322168] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.322531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.322560] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.322938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.322968] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.323324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.323352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.323726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.323755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.324122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.324149] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.324522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.324552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.324846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.324875] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.325235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.325264] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.325634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.325662] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.326048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.326078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.326349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.326379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.326721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.326752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.327112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.327141] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.327352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.327382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.327675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.327715] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.328135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.328163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.328528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.328555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.328916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.328946] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.329308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.329344] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.329684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.329724] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.330068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.330105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.330434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.330462] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.330816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.330845] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.331085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.331117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.331510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.331540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.331869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.331898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.332270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.332297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.332675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.332714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.535 [2024-10-17 17:50:06.333068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.535 [2024-10-17 17:50:06.333095] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.535 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.333462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.333491] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.333863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.333893] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.334248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.334276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.334618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.334646] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.335015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.335045] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.335410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.335438] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.335834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.335864] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.336288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.336316] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.336686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.336738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.337100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.337127] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.337501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.337530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.337890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.337919] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.338281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.338310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.338685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.338724] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.339083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.339111] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.339478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.339506] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.339859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.339894] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.340257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.340287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.340665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.340703] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.341069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.341096] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.341521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.341548] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.341890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.341918] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.342290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.342317] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.342686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.342729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.342982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.343010] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.343349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.343378] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.343614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.343643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.344032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.344061] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.344425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.344454] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.344817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.344846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.345215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.345244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.345605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.345633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.345997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.346027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.346436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.346464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.346909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.346938] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.347317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.347346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.347583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.347616] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.347998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.348026] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.348387] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.348415] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.348750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.536 [2024-10-17 17:50:06.348781] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.536 qpair failed and we were unable to recover it. 00:28:58.536 [2024-10-17 17:50:06.349142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.349170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.349539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.349568] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.349914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.349944] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.350312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.350340] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.350713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.350742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.351114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.351142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.351495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.351524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.351770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.351799] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.352167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.352195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.352463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.352490] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.352845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.352874] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.353244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.353273] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.353632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.353659] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.354034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.354065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.354427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.354455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.354824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.354853] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.355224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.355258] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.355617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.355644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.355892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.355924] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.356217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.356245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.356462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.356493] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.356853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.356883] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.357259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.357287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.357664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.357703] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.358059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.358087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.358458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.358487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.358841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.358869] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.359232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.359260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.359633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.359660] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.360032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.360061] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.360421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.360451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.360810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.360841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.361219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.361247] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.361494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.361525] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.361917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.361946] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.362298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.362325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.362589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.362618] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.362954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.362985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.363317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.363347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.363714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.363744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.537 [2024-10-17 17:50:06.364003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.537 [2024-10-17 17:50:06.364030] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.537 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.364382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.364410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.364777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.364808] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.365056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.365088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.365441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.365470] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.367289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.367355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.367796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.367835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.368235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.368265] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.368627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.368655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.369043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.369073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.369436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.369464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.369827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.369858] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.370192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.370221] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.370601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.370629] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.371060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.371090] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.371424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.371452] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.371666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.371718] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.372097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.372126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.372486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.372516] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.372965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.372995] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.373350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.373379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.373750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.373780] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.374189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.374219] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.374460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.374492] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.374851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.374880] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.375243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.375271] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.375625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.375652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.376017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.376047] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.376408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.376438] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.376812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.376842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.377228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.377256] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.377717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.377748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.378004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.378031] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.378304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.378332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.378588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.378620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.378988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.379019] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.379475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.379503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.379760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.379789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.380194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.380222] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.380579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.380607] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.380962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.538 [2024-10-17 17:50:06.380990] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.538 qpair failed and we were unable to recover it. 00:28:58.538 [2024-10-17 17:50:06.381357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.381386] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.381763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.381795] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.382145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.382173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.382528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.382557] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.382814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.382843] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.383080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.383111] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.383464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.383494] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.383828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.383858] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.384231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.384259] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.384488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.384518] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.384900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.384928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.385341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.385370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.385728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.385758] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.386161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.386190] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.386443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.386474] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.386848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.386888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.387123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.387156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.387500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.387527] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.387898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.387929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.388333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.388361] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.388726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.388754] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.389095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.389123] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.389367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.389403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.389760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.389790] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.390186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.390216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.390653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.390682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.391112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.391141] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.391512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.391540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.391797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.391826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.392216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.392247] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.392603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.392633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.392984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.393014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.393384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.393414] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.393658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.393687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.394068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.394098] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.394461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.394491] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.539 qpair failed and we were unable to recover it. 00:28:58.539 [2024-10-17 17:50:06.394742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.539 [2024-10-17 17:50:06.394777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.395141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.395169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.395527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.395556] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.395957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.395988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.396357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.396385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.396756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.396789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.397158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.397188] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.397551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.397581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.397958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.397987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.398354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.398384] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.398747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.398777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.399128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.399156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.399513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.399542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.399885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.399914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.400266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.400295] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.400671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.400711] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.401074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.401102] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.401480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.401510] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.401852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.401882] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.402236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.402271] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.402623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.402652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.403042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.403073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.403439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.403478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.403809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.403840] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.404111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.404140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.404482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.404511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.404771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.404800] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.405171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.405201] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.405472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.405501] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.405850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.405881] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.406128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.406159] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.406507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.406537] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.406785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.406817] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.407198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.407228] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.407583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.407612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.407937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.407970] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.408360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.408389] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.408712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.408743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.409120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.409150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.409513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.409542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.409908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.540 [2024-10-17 17:50:06.409937] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.540 qpair failed and we were unable to recover it. 00:28:58.540 [2024-10-17 17:50:06.410207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.410236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.410514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.410543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.410713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.410741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.411100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.411129] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.411489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.411519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.411938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.411970] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.412299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.412327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.412751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.412782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.413151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.413180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.413420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.413448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.413883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.413913] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.414267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.414296] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.414686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.414730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.415089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.415118] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.415476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.415505] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.415861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.415891] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.416235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.416266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.416436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.416466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.416749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.416787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.417113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.417144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.417502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.417532] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.417943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.417975] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.418392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.418421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.418791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.418823] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.419214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.419244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.419557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.419585] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.419857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.419887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.420117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.420149] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.420405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.420435] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.420854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.420884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.421249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.421279] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.421645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.421673] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.422055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.422087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.422443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.422471] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.422789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.422819] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.423192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.423220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.423571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.423602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.423991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.424023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.424253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.424281] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.424517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.424547] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.541 [2024-10-17 17:50:06.424793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.541 [2024-10-17 17:50:06.424823] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.541 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.425224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.425252] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.425626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.425656] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.426046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.426077] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.426435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.426463] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.426830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.426862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.427057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.427085] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.427505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.427534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.427953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.427984] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.428358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.428392] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.428637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.428666] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.429122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.429152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.542 [2024-10-17 17:50:06.429595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.542 [2024-10-17 17:50:06.429625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.542 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.429955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.429986] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.430366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.430399] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.430636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.430669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.431121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.431151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.431506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.431535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.431886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.431925] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.432310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.432339] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.432715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.432747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.433110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.433139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.433501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.433531] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.433893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.433922] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.434282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.434311] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.434658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.434687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.434928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.434959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.435309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.435338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.435573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.435601] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.435862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.435894] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.436275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.436305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.436665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.436706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.437101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.437131] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.437565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.437593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.437943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.437973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.438327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.438356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.438533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.438561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.438928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.438958] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.439316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.439345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.439734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.439766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.440029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.440058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.440454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.440484] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.440830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.440861] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.441214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.441242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.441604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.441633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.442038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.442069] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.442327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.442355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.442751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.442782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.443150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.443178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.443540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.443569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.443823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.443854] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.815 [2024-10-17 17:50:06.444208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.815 [2024-10-17 17:50:06.444236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.815 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.444569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.444600] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.444945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.444976] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.445349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.445379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.445751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.445780] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.446151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.446179] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.446550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.446579] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.446925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.446961] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.447326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.447356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.447731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.447763] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.448170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.448202] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.448574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.448605] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.448862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.448892] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.449246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.449274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.449652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.449681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.449955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.449985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.450374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.450403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.450680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.450722] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.450993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.451023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.451314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.451341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.451716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.451753] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.452136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.452167] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.452542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.452570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.452876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.452907] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.453248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.453276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.453658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.453688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.454084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.454113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.454391] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.454419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.454777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.454806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.455150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.455178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.455324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.455356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.455739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.455770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.456144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.456172] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.456586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.456613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.457010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.457041] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.457402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.457430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.457812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.457841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.458214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.458245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.458596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.458625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.458806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.458835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.816 qpair failed and we were unable to recover it. 00:28:58.816 [2024-10-17 17:50:06.459221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.816 [2024-10-17 17:50:06.459249] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.459585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.459612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.460015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.460045] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.460450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.460479] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.460736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.460765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.461020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.461048] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.461423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.461451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.461857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.461892] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.462254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.462282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.462546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.462575] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.462887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.462915] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.463276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.463304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.463688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.463743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.464008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.464035] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.464388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.464416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.464865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.464895] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.465208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.465237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.465574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.465603] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.465958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.465987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.466281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.466309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.466709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.466739] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.467109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.467139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.467492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.467521] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.467911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.467941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.468335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.468363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.468727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.468757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.469143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.469171] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.469565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.469593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.469918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.469947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.470228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.470256] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.470676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.470714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.471123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.471151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.471514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.471550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.471906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.471936] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.472187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.472216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.472556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.472584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.472858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.472888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.473269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.473297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.473664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.473702] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.473926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.473954] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.474312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.817 [2024-10-17 17:50:06.474341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.817 qpair failed and we were unable to recover it. 00:28:58.817 [2024-10-17 17:50:06.474659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.474687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.475112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.475140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.475590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.475617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.476071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.476100] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.476385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.476414] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.476772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.476802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.477032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.477069] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.477451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.477480] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.477836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.477865] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.478248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.478276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.478642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.478671] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.479045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.479074] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.479317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.479347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.479596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.479625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.480056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.480086] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.480441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.480470] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.480714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.480744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.481078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.481106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.481487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.481516] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.481933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.481962] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.482374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.482403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.482763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.482794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.483051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.483078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.483529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.483558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.483784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.483813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.484185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.484213] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.484569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.484606] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.484875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.484904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.485286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.485314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.485676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.485736] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.486125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.486154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.486398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.486430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.486773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.486802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.487157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.487186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.487547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.487575] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.487923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.487954] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.488180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.488210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.488589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.488618] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.488848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.488880] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.489246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.489274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.818 qpair failed and we were unable to recover it. 00:28:58.818 [2024-10-17 17:50:06.489630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.818 [2024-10-17 17:50:06.489660] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.489943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.489973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.490374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.490403] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.490659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.490688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.490951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.490984] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.491363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.491392] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.491801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.491836] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.492276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.492305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.492672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.492713] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.493076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.493104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.493469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.493497] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.493758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.493787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.494167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.494195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.494556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.494584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.494933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.494963] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.495317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.495346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.495600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.495628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.496001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.496032] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.496390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.496419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.496787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.496818] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.497170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.497201] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.497530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.497558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.497814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.497848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.498225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.498253] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.498597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.498625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.498865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.498898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.499241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.499272] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.499629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.499657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.500059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.500089] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.500449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.500478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.500844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.500873] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.501246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.501276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.501523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.501555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.501804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.819 [2024-10-17 17:50:06.501837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.819 qpair failed and we were unable to recover it. 00:28:58.819 [2024-10-17 17:50:06.502213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.502241] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.502616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.502645] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.503115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.503145] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.503483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.503512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.503763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.503796] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.504197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.504225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.504591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.504619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.504986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.505015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.505379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.505408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.505775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.505807] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.506198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.506226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.506586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.506614] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.506861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.506897] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.507260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.507290] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.507664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.507715] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.507976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.508008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.508276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.508307] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.508684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.508727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.508977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.509005] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.509358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.509387] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.509751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.509782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.510145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.510175] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.510572] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.510600] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.510945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.510975] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.511341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.511370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.511749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.511778] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.512139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.512176] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.512510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.512539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.512874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.512903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.513279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.513307] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.513536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.513567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.513942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.513972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.514312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.514344] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.514689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.514729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.515088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.515117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.515494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.515522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.515922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.515953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.516182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.516213] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.516575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.516606] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.516982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.517014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.517374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.820 [2024-10-17 17:50:06.517402] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.820 qpair failed and we were unable to recover it. 00:28:58.820 [2024-10-17 17:50:06.517766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.517795] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.518181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.518209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.518575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.518603] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.518945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.518974] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.519338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.519367] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.519711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.519741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.520011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.520038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.520414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.520442] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.520769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.520797] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.521195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.521223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.521588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.521617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.521968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.522003] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.522367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.522396] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.522796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.522825] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.523171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.523208] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.523581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.523610] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.523960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.523989] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.524332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.524360] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.524726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.524757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.525125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.525152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.525521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.525550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.525890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.525921] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.526301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.526328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.526709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.526738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.527001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.527033] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.527329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.527357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.527778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.527807] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.528052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.528084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.528454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.528483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.528850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.528881] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.529264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.529292] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.529649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.529677] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.530059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.530087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.530325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.530354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.530714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.530745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.530983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.531011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.531400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.531429] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.531807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.531836] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.532075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.532113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.532496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.532526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.821 [2024-10-17 17:50:06.532899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.821 [2024-10-17 17:50:06.532931] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.821 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.533291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.533319] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.533686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.533726] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.533979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.534010] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.534379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.534409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.534777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.534809] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.535177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.535205] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.535574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.535602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.535852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.535884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.536246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.536274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.536647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.536676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.537046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.537077] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.537306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.537337] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.537634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.537663] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.538049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.538079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.538443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.538471] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.538836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.538866] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.539239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.539268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.539633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.539662] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.540062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.540093] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.540429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.540458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.540832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.540862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.541300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.541328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.541660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.541689] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.541927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.541955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.542293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.542324] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.542758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.542788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.543151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.543178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.543538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.543568] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.543821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.543852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.544222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.544252] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.544653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.544681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.545030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.545058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.545409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.545438] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.545808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.545839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.546216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.546245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.546680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.546723] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.547079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.547108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.547483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.547517] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.547876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.547906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.548259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.822 [2024-10-17 17:50:06.548287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.822 qpair failed and we were unable to recover it. 00:28:58.822 [2024-10-17 17:50:06.548666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.548706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.549066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.549096] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.549432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.549460] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.549726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.549755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.550095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.550122] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.550299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.550331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.550716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.550746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.551158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.551186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.551549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.551577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.551827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.551860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.552226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.552254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.552616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.552645] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.553006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.553036] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.553403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.553432] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.553810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.553838] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.554234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.554264] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.554611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.554640] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.555034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.555065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.555430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.555458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.555821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.555852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.556184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.556212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.556583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.556611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.556854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.556887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.557267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.557296] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.557709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.557741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.558100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.558128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.558496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.558524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.558899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.558928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.559287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.559315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.559725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.559756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.560111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.560139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.560485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.560514] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.560898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.560929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.561298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.561327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.561577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.561609] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.823 qpair failed and we were unable to recover it. 00:28:58.823 [2024-10-17 17:50:06.562004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.823 [2024-10-17 17:50:06.562034] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.562412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.562440] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.562800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.562834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.563201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.563230] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.563591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.563619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.563988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.564019] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.564359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.564388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.564762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.564792] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.565124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.565153] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.565496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.565524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.565898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.565928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.566293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.566323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.566687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.566727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.567079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.567108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.567474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.567501] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.567871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.567901] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.568177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.568206] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.568562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.568591] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.568977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.569008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.569438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.569466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.569798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.569828] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.570182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.570210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.570580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.570609] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.570974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.571003] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.571370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.571397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.571770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.571801] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.572133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.572161] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.572526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.572553] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.572930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.572960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.573306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.573335] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.573719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.573750] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.574105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.574133] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.574489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.574518] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.574915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.574944] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.575298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.575327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.575710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.575742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.576136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.576164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.576526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.576554] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.576925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.576956] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.577318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.577345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.577715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.824 [2024-10-17 17:50:06.577745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.824 qpair failed and we were unable to recover it. 00:28:58.824 [2024-10-17 17:50:06.578124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.578153] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.578515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.578549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.578920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.578949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.579321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.579349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.579722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.579754] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.580013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.580042] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.580326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.580356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.580751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.580781] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.581037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.581067] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.581452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.581481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.581836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.581867] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.582238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.582269] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.582635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.582666] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.583038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.583067] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.583428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.583457] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.583820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.583850] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.584196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.584225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.584615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.584644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.585016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.585046] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.585384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.585412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.585771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.585801] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.586170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.586199] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.586443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.586475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.586835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.586867] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.587136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.587164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.587511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.587539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.587895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.587925] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.588303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.588332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.588711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.588741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.589119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.589148] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.589454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.589483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.589824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.589853] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.590182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.590212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.590562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.590590] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.590983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.591012] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.591345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.591373] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.591612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.591640] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.592004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.592034] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.592393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.592421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.592789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.592819] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.825 qpair failed and we were unable to recover it. 00:28:58.825 [2024-10-17 17:50:06.593165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.825 [2024-10-17 17:50:06.593193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.593550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.593589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.593936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.593965] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.594205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.594237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.594447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.594475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.594892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.594921] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.595288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.595315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.595683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.595735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.596088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.596116] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.596474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.596503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.596927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.596957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.597218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.597246] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.597598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.597627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.597945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.597974] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.598355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.598382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.598644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.598673] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.599029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.599057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.599417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.599445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.599728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.599757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.600124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.600152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.600565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.600593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.600953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.600983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.601355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.601382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.601753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.601783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.602142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.602169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.602521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.602550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.602998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.603027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.603390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.603418] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.603822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.603852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.604219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.604248] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.604614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.604643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.605020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.605050] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.605289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.605321] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.605727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.605757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.606099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.606127] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.606496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.606524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.606796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.606825] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.607207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.607235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.607615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.607644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.608111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.608139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.608496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.826 [2024-10-17 17:50:06.608524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.826 qpair failed and we were unable to recover it. 00:28:58.826 [2024-10-17 17:50:06.608891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.608926] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.609331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.609359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.609715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.609744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.610138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.610166] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.610542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.610569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.610931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.610961] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.611330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.611358] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.611732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.611761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.612111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.612139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.612509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.612537] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.612977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.613006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.613352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.613379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.613751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.613783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.614172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.614200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.614452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.614481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.614843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.614874] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.615259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.615287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.615544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.615572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.615816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.615846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.616224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.616254] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.616520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.616549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.616933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.616963] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.617325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.617353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.617727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.617757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.618116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.618144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.618581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.618611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.618939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.618969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.619206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.619238] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.619604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.619633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.620007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.620037] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.620419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.620447] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.620745] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.620776] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.621141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.621170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.621546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.621574] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.621957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.621988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.622343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.827 [2024-10-17 17:50:06.622370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.827 qpair failed and we were unable to recover it. 00:28:58.827 [2024-10-17 17:50:06.622738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.622769] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.623151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.623180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.623545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.623573] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.623971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.623999] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.624360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.624394] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.624755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.624785] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.625157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.625185] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.625542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.625571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.625934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.625963] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.626332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.626360] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.626727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.626757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.627110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.627139] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.627497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.627527] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.627874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.627903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.628266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.628294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.628658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.628686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.629027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.629057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.629298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.629328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.629730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.629762] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.630165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.630194] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.630556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.630584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.630919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.630949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.631310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.631338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.631712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.631742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.632093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.632124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.632466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.632494] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.632888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.632919] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.633154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.633183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.633482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.633512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.633925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.633957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.634353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.634382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.634746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.634777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.635149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.635178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.635538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.635566] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.635826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.635855] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.636285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.636315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.636648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.636676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.638508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.638567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.639014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.639052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.640688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.640769] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.828 [2024-10-17 17:50:06.641086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.828 [2024-10-17 17:50:06.641117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.828 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.642891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.642943] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.643327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.643359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.643726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.643757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.643991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.644040] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.644413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.644443] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.644818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.644849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.645212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.645242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.645605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.645635] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.645980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.646012] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.646353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.646384] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.646644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.646676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.646959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.646988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.647369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.647398] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.647749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.647780] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.648179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.648207] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.648446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.648475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.648729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.648761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.649193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.649225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.649582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.649611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.650007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.650041] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.650401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.650430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.650887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.650920] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.651312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.651341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.651749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.651781] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.652033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.652069] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.652428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.652458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.652846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.652877] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.653235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.653265] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.653514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.653545] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.653933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.653964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.654357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.654388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.654764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.654798] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.655206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.655238] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.655645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.655674] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.656072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.656101] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.656441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.656471] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.656808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.656839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.657214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.657245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.657587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.657618] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.657981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.829 [2024-10-17 17:50:06.658011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.829 qpair failed and we were unable to recover it. 00:28:58.829 [2024-10-17 17:50:06.658279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.658308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.658718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.658748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.659169] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.659200] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.659558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.659593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.659943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.659973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.660350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.660380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.660762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.660793] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.661154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.661184] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.661630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.661661] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.661915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.661949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.662319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.662350] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.662732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.662763] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.663131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.663159] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.663576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.663604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.663930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.663961] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.664308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.664336] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.664596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.664625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.665041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.665072] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.665424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.665453] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.665813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.665845] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.666134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.666163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.666567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.666596] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.666968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.666997] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.667348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.667376] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.667637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.667665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.667934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.667964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.668343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.668373] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.668629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.668659] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.669108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.669138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.669389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.669419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.669802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.669834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.670085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.670117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.670456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.670487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.670855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.670887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.671303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.671332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.671723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.671753] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.672121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.672150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.672521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.672550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.672997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.673028] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.673382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.673412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.830 [2024-10-17 17:50:06.673744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.830 [2024-10-17 17:50:06.673774] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.830 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.674137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.674165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.674538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.674568] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.674800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.674839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.675205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.675236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.675601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.675629] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.675951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.675982] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.676319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.676349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.676613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.676642] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.677040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.677070] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.677427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.677457] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.677794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.677825] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.678212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.678241] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.678615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.678643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.678932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.678962] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.679330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.679359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.679748] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.679777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.680164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.680194] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.680540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.680569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.680837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.680870] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.681117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.681148] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.681528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.681559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.681976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.682007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.682348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.682376] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.682746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.682775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.683067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.683095] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.683474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.683503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.683870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.683900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.684237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.684266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.684638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.684669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.685027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.685058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.685420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.685449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.685901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.685932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.686264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.686294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.686664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.686706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.687126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.687154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.687362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.687391] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.687648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.687682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.688058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.688088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.688452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.688480] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.688877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.831 [2024-10-17 17:50:06.688908] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.831 qpair failed and we were unable to recover it. 00:28:58.831 [2024-10-17 17:50:06.689272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.689300] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.689661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.689690] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.689950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.689989] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.690369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.690396] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.690761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.690792] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.691175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.691204] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.691588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.691617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.691957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.691987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.692239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.692267] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.692675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.692730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.693124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.693153] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.693530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.693558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.693978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.694007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.694375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.694404] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.694799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.694829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.695205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.695233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.695616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.695646] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.695944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.695973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.696320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.696348] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.696798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.696827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.697217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.697245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.697622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.697651] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.698036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.698066] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.698425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.698453] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.698821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.698852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.699179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.699206] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.699539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.699568] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.699809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.699841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.700245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.700274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.700659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.700688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.701079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.701107] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.701486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.701513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.701943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.701973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.702335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.702364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.702634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.702663] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.703017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.832 [2024-10-17 17:50:06.703046] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.832 qpair failed and we were unable to recover it. 00:28:58.832 [2024-10-17 17:50:06.703401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.703430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.703816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.703846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.704259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.704287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.704627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.704655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.705081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.705111] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.705457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.705485] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.705866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.705904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.706260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.706291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.706638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.706665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.707074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.707104] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.707458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.707487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.707741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.707771] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.708207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.708235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.708498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.708526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.708895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.708924] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.709329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.709357] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.709736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.709766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.710113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.710141] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.710521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.710549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.710972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.711002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.711378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.711407] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.711775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.711806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.712065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.712094] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.712465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.712493] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.712776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.712805] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.713157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.713184] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.713553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.713582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.713984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.714013] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.714246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.714277] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.714700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.714730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.715072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.715101] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.715589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.715617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.716000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.716030] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.716409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.716439] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.716785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.716815] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.717184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.717212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.717582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.717610] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.718015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.718044] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.718437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.718466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.718717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.718747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.833 [2024-10-17 17:50:06.719102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.833 [2024-10-17 17:50:06.719130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.833 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.719436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.719464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.719844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.719873] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.720122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.720151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.720415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.720444] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.720715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.720748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.721207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.721241] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.721583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.721613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.721952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.721981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:58.834 [2024-10-17 17:50:06.722342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:58.834 [2024-10-17 17:50:06.722370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:58.834 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.722734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.722765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.723125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.723156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.723523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.723551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.723811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.723842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.724215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.724242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.724601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.724630] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.724890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.724924] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.725307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.725338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.725712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.725742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.726147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.726174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.726542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.726572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.726925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.726955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.106 [2024-10-17 17:50:06.727280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.106 [2024-10-17 17:50:06.727309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.106 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.727520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.727548] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.727920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.727951] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.728314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.728342] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.728689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.728730] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.728990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.729018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.729357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.729385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.729764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.729793] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.730178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.730205] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.730622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.730652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.731029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.731058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.731425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.731455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.731738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.731769] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.732142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.732171] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.732538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.732568] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.732917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.732946] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.733317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.733345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.733599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.733627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.733976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.734005] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.734368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.734397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.734764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.734795] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.735188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.735217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.735560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.735588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.736004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.736034] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.736392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.736420] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.736791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.736821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.737193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.737223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.737595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.737622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.737985] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.738015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.738376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.738405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.738772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.738801] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.739010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.739041] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.739403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.739433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.739733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.739765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.740121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.740149] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.740526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.740555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.740992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.741021] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.741383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.741410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.741734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.741764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.742129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.742157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.107 [2024-10-17 17:50:06.742558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.107 [2024-10-17 17:50:06.742585] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.107 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.742940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.742969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.743333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.743361] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.743725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.743756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.744154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.744183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.744554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.744582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.744924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.744952] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.745210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.745237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.745595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.745623] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.745986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.746015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.746380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.746409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.746752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.746787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.747160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.747189] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.747554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.747582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.747839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.747868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.748204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.748231] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.748595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.748624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.749011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.749041] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.749407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.749434] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.749800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.749829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.749971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.750002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.750354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.750382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.750762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.750792] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.751143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.751172] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.751552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.751580] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.751962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.751992] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.752354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.752383] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.752749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.752779] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.753149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.753178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.753542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.753569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.753999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.754028] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.754283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.754314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.754751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.754780] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.755178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.755206] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.755544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.755572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.755947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.755978] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.756331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.756359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.756781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.756810] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.757184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.757212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.757574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.757604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.757997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.108 [2024-10-17 17:50:06.758027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.108 qpair failed and we were unable to recover it. 00:28:59.108 [2024-10-17 17:50:06.758393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.758421] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.758864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.758893] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.759247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.759275] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.759714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.759746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.760130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.760158] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.760522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.760549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.760916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.760944] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.761342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.761370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.761734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.761763] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.762111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.762140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.762508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.762542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.762648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.762678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.763107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.763136] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.763507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.763535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.763872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.763903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.764267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.764297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.764656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.764685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.765054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.765084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.765458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.765486] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.765832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.765860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.766233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.766262] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.766624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.766653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.766908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.766941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.767290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.767318] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.767706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.767736] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.768132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.768159] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.768510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.768540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.768884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.768914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.769160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.769191] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.769548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.769578] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.769923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.769953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.770325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.770353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.770587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.770618] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.770898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.770928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.771294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.771322] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.771690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.771729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.772072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.772100] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.772458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.772486] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.772732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.772765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.773127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.109 [2024-10-17 17:50:06.773157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.109 qpair failed and we were unable to recover it. 00:28:59.109 [2024-10-17 17:50:06.773523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.773552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.773918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.773947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.774310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.774338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.774568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.774596] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.775042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.775071] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.775446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.775476] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.775826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.775855] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.776231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.776260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.776609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.776637] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.777019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.777049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.777416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.777451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.777799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.777830] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.778211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.778239] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.778607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.778635] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.779007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.779038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.779398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.779428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.779795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.779826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.780100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.780128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.780518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.780546] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.780919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.780949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.781324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.781353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.781716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.781746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.782127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.782156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.782398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.782426] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.782778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.782809] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.783216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.783244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.783610] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.783638] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.783997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.784031] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.784266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.784297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.784547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.784577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.784824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.784857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.785163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.785191] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.785549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.785578] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.785953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.785983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.786345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.786373] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.110 [2024-10-17 17:50:06.786742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.110 [2024-10-17 17:50:06.786771] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.110 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.787155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.787183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.787544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.787573] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.787834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.787863] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.788241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.788269] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.788630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.788657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.789044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.789073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.789431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.789459] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.789829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.789858] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.790295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.790323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.790658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.790687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.791048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.791076] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.791436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.791464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.791830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.791859] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.792150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.792178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.792609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.792648] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.793028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.793058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.793417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.793445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.793711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.793742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.794126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.794154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.794516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.794543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.794900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.794930] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.795292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.795320] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.795687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.795727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.796063] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.796091] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.796352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.796380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.796735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.796764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.797129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.797157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.797415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.797447] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.797807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.797837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.798173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.798201] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.798565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.798593] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.798938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.798968] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.799217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.799250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.799584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.799613] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.799979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.800009] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.800309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.800337] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.800599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.800631] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.800990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.801020] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.111 [2024-10-17 17:50:06.801381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.111 [2024-10-17 17:50:06.801410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.111 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.801662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.801702] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.802059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.802088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.802454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.802483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.802841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.802872] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.803237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.803265] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.803626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.803654] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.804017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.804047] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.804408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.804436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.804812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.804841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.805193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.805221] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.805490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.805519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.805912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.805941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.806310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.806338] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.806591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.806622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.807036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.807066] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.807428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.807462] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.807828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.807857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.808117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.808149] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.808510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.808539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.808775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.808807] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.809182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.809209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.809583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.809611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.809870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.809902] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.810260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.810288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.810649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.810677] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.810972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.811002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.811335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.811363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.811714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.811744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.812094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.812122] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.812480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.812508] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.812886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.812916] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.813280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.813308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.813540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.813567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.813943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.813971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.814405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.814433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.814798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.814826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.815198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.815226] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.815591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.815619] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.815982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.816011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.816378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.112 [2024-10-17 17:50:06.816406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.112 qpair failed and we were unable to recover it. 00:28:59.112 [2024-10-17 17:50:06.816775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.816804] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.817190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.817218] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.817551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.817580] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.817974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.818002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.818236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.818268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.818516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.818548] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.818922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.818951] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.819215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.819243] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.819605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.819633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.819994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.820024] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.820387] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.820415] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.820648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.820679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.821051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.821080] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.821310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.821340] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.821734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.821764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.822052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.822088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.822438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.822466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.822835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.822864] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.823239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.823266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.823571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.823599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.823942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.823971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.824336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.824364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.824711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.824740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.825089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.825116] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.825235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.825266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.825641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.825670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.826040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.826070] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.826446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.826473] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.826836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.826866] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.827222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.827250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.827620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.827648] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.828015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.828044] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.828400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.828427] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.828792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.828821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.829068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.829100] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.829472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.829500] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.829869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.829899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.830269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.830297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.830658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.830686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.831099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.831128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.113 [2024-10-17 17:50:06.831476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.113 [2024-10-17 17:50:06.831503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.113 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.831871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.831901] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.832154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.832182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.832537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.832565] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.832928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.832957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.833357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.833385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.833748] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.833777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.834150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.834178] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.834546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.834573] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.834962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.834992] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.835355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.835383] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.835749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.835779] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.836148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.836176] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.836609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.836637] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.836967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.836996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.837361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.837395] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.837748] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.837776] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.838146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.838174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.838551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.838579] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.838931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.838959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.839196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.839228] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.839455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.839483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.839860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.839890] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.840254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.840282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.840711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.840740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.841103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.841130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.841494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.841523] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.841875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.841904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.842172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.842199] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.842556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.842585] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.842929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.842959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.843319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.843347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.843765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.843794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.844159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.844186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.844537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.844564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.844914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.844943] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.845204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.845232] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.845584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.845612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.845993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.846021] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.846383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.846411] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.114 qpair failed and we were unable to recover it. 00:28:59.114 [2024-10-17 17:50:06.846774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.114 [2024-10-17 17:50:06.846803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.847171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.847199] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.847508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.847538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.847910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.847940] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.848190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.848218] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.848556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.848584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.848797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.848829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.849183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.849210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.849568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.849595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.849997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.850025] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.850390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.850418] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.850780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.850809] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.851179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.851207] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.851571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.851599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.851827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.851857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.852255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.852289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.852646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.852675] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.853052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.853082] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.853436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.853464] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.853870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.853899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.854255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.854283] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.854655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.854682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.855069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.855098] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.855464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.855492] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.855855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.855885] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.856248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.856276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.856642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.856669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.857029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.857058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.857439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.857467] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.857769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.857821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.858164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.858192] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.858558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.858587] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.858961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.858990] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.859351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.859379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.859739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.859768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.115 [2024-10-17 17:50:06.860129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.115 [2024-10-17 17:50:06.860156] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.115 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.860523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.860551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.860925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.860953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.861313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.861341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.861711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.861740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.861974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.862006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.862353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.862382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.862751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.862782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.863232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.863260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.863593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.863620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.864025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.864055] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.864399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.864428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.864790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.864819] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.865184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.865212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.865585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.865612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.865986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.866015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.866378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.866406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.866754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.866783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.867208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.867236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.867494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.867522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.867871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.867912] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.868249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.868277] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.868637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.868665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.868916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.868945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.869299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.869326] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.869729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.869757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.870032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.870061] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.870312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.870342] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.870690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.870737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.871146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.871173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.871521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.871549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.872521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.872555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.873003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.873035] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.873413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.873441] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.873809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.873839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.874207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.874235] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.874483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.874511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.874944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.874973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.875330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.875359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.875622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.875649] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.876009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.116 [2024-10-17 17:50:06.876038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.116 qpair failed and we were unable to recover it. 00:28:59.116 [2024-10-17 17:50:06.876396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.876425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.876794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.876824] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.877196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.877223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.877624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.877652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.878016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.878045] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.878411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.878438] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.878801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.878832] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.879205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.879232] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.879596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.879625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.879867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.879899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.880263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.880291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.880651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.880679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.881109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.881137] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.881489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.881517] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.881830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.881860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.882241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.882268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.882641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.882670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.883080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.883110] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.883353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.883381] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.883775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.883810] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.884166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.884194] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.884557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.884585] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.884931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.884960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.885321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.885349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.885720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.885751] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.886110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.886138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.886558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.886586] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.886956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.886985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.887400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.887428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.887791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.887820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.888190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.888217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.888588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.888616] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.888982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.889012] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.889351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.889380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.889746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.889776] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.890181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.890209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.890561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.890588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.890960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.890989] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.891363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.891391] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.891757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.891786] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.117 [2024-10-17 17:50:06.892166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.117 [2024-10-17 17:50:06.892193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.117 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.892561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.892589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.892935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.892963] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.893341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.893369] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.893724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.893754] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.894141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.894169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.894540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.894568] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.894917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.894946] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.895307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.895335] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.895707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.895737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.895951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.895981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.896317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.896345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.896706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.896736] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.897035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.897063] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.897421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.897449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.897800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.897829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.898207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.898234] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.898598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.898626] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.898972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.899002] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.899360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.899394] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.899757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.899787] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.900166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.900193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.900558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.900586] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.900930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.900959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.901298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.901327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.901689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.901747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.902107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.902135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.902478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.902506] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.902871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.902901] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.903256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.903285] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.903651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.903679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.904075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.904103] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.904501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.904529] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.904853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.904882] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.905243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.905272] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.905635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.905665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.906003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.906032] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.906395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.906422] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.906778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.906809] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.907181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.907209] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.907569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.118 [2024-10-17 17:50:06.907598] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.118 qpair failed and we were unable to recover it. 00:28:59.118 [2024-10-17 17:50:06.907971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.908000] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.908357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.908385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.908788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.908818] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.909217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.909244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.909603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.909631] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.909978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.910009] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.910365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.910393] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.910757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.910788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.911151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.911179] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.911541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.911569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.911921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.911950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.912303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.912331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.912704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.912733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.913034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.913061] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.913419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.913447] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.913809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.913838] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.914214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.914241] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.914603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.914630] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.914845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.914881] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.915250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.915278] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.915642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.915671] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.916083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.916113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.916472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.916499] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.916725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.916759] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.917118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.917146] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.917504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.917532] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.917770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.917802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.918250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.918278] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.918638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.918666] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.919074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.919103] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.919472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.919500] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.919865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.919894] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.920254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.920283] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.920644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.920672] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.921045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.921073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.119 [2024-10-17 17:50:06.921411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.119 [2024-10-17 17:50:06.921439] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.119 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.921812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.921840] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.922221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.922248] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.922499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.922531] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.922916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.922945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.923279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.923308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.923661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.923702] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.924065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.924093] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.924535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.924563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.924936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.924966] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.925365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.925393] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.925687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.925727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.926100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.926128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.926495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.926522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.926744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.926774] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.927135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.927163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.927528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.927557] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.927921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.927951] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.928320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.928347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.928716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.928746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.929111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.929140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.929502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.929530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.929903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.929932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.930307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.930334] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.930703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.930734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.930989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.931021] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.931351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.931378] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.931749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.931779] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.932135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.932162] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.932530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.932558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.932913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.932943] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.933191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.933223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.933457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.933488] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.933853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.933883] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.934240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.934267] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.934627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.934654] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.935013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.935042] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.935421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.935449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.935756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.935785] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.936162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.936190] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.936544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.936572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.120 qpair failed and we were unable to recover it. 00:28:59.120 [2024-10-17 17:50:06.936951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.120 [2024-10-17 17:50:06.936982] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.937280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.937308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.937548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.937579] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.937936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.937966] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.938397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.938425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.938756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.938786] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.939165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.939193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.939553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.939580] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.939935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.939965] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.940347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.940382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.940739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.940770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.941129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.941157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.941411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.941441] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.941794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.941824] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.942183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.942211] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.942463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.942495] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.942759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.942788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.943137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.943166] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.943533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.943560] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.943793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.943825] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.944191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.944219] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.944583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.944611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.944984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.945014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.945386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.945414] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.945778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.945808] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.946208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.946236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.946590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.946618] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.946879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.946912] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.947167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.947195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.947480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.947508] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.947773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.947802] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.948147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.948175] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.948538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.948566] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.948869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.948899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.949268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.949295] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.949536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.949567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.949874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.949904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.950253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.950280] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.950648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.950676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.951045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.951074] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.121 [2024-10-17 17:50:06.951433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.121 [2024-10-17 17:50:06.951460] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.121 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.951833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.951861] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.952231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.952259] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.952622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.952650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.953089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.953119] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.953493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.953522] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.953906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.953937] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.954239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.954267] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.954529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.954556] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.954934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.954970] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.955373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.955401] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.955760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.955789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.956163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.956191] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.956557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.956584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.956926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.956955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.957318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.957345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.957719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.957749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.958110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.958138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.958509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.958536] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.958879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.958908] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.959265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.959293] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.959522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.959553] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.959932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.959964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.960300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.960328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.960689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.960731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.961126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.961154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.961522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.961550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.961936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.961967] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.962338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.962366] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.962732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.962761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.963097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.963125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.963505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.963534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.963871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.963899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.964273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.964301] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.964711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.964740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.965097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.965124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.965531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.965561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.965992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.966022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.966380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.966408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.966784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.966813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.967180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.122 [2024-10-17 17:50:06.967208] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.122 qpair failed and we were unable to recover it. 00:28:59.122 [2024-10-17 17:50:06.967582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.967610] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.967976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.968004] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.968372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.968402] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.968760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.968789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.969188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.969216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.969580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.969609] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.969980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.970011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.970376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.970404] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.970680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.970725] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.970986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.971020] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.971374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.971402] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.971650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.971682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.972076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.972105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.972357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.972386] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.972737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.972766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.973154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.973183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.973553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.973580] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.973933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.973964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.974321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.974349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.974711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.974741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.975144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.975172] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.975535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.975564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.975947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.975976] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.976208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.976239] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.976616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.976644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.977018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.977047] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.977416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.977445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.977809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.977839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.978109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.978137] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.978478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.978509] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.978874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.978903] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.979258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.979286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.979655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.979682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.980059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.980087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.980449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.980477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.123 qpair failed and we were unable to recover it. 00:28:59.123 [2024-10-17 17:50:06.980877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.123 [2024-10-17 17:50:06.980907] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.981270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.981298] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.981509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.981541] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.981918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.981949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.982311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.982339] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.982711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.982741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.983087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.983115] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.983476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.983504] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.983880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.983909] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.984272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.984302] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.984705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.984737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.985085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.985113] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.985484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.985511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.985861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.985899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.986263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.986291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.986651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.986679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.987049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.987079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.987441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.987468] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.987747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.987777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.988154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.988181] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.988560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.988588] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.988929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.988959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.989319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.989347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.989706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.989737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.989967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.989995] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.990350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.990379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.990792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.990822] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.991167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.991195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.991436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.991465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.991809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.991838] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.992216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.992244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.992613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.992641] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.993081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.993110] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.993351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.993384] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.993727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.993758] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.994099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.994127] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.994392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.994422] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.994771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.994803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.995170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.995199] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.995634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.995662] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.124 [2024-10-17 17:50:06.996043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.124 [2024-10-17 17:50:06.996073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.124 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.996431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.996460] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.996832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.996861] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.997240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.997268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.997526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.997554] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.997920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.997949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.998315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.998343] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.998591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.998621] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.998975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.999006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.999371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.999399] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:06.999763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:06.999791] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.000153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.000182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.000543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.000572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.000830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.000865] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.001221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.001250] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.001621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.001651] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.001991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.002023] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.002366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.002395] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.002754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.002786] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.003154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.003182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.003434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.003467] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.003860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.003890] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.004251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.004280] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.004654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.004683] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.005064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.005093] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.005458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.005487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.005852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.005882] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.006235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.006265] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.006642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.006671] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.007034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.007065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.007369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.007398] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.007767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.007798] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.008160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.008188] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.008552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.008580] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.008954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.008985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.009352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.009380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.009701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.009731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.010133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.010162] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.010521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.010550] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.010812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.010841] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.011221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.125 [2024-10-17 17:50:07.011251] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.125 qpair failed and we were unable to recover it. 00:28:59.125 [2024-10-17 17:50:07.011611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.126 [2024-10-17 17:50:07.011639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.126 qpair failed and we were unable to recover it. 00:28:59.126 [2024-10-17 17:50:07.012000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.126 [2024-10-17 17:50:07.012031] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.126 qpair failed and we were unable to recover it. 00:28:59.126 [2024-10-17 17:50:07.012388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.126 [2024-10-17 17:50:07.012416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.126 qpair failed and we were unable to recover it. 00:28:59.126 [2024-10-17 17:50:07.012705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.126 [2024-10-17 17:50:07.012736] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.126 qpair failed and we were unable to recover it. 00:28:59.126 [2024-10-17 17:50:07.013104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.126 [2024-10-17 17:50:07.013132] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.126 qpair failed and we were unable to recover it. 00:28:59.126 [2024-10-17 17:50:07.013440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.126 [2024-10-17 17:50:07.013468] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.126 qpair failed and we were unable to recover it. 00:28:59.126 [2024-10-17 17:50:07.013837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.126 [2024-10-17 17:50:07.013868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.126 qpair failed and we were unable to recover it. 00:28:59.397 [2024-10-17 17:50:07.014226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.397 [2024-10-17 17:50:07.014255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.397 qpair failed and we were unable to recover it. 00:28:59.397 [2024-10-17 17:50:07.014614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.397 [2024-10-17 17:50:07.014644] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.397 qpair failed and we were unable to recover it. 00:28:59.397 [2024-10-17 17:50:07.015001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.397 [2024-10-17 17:50:07.015031] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.397 qpair failed and we were unable to recover it. 00:28:59.397 [2024-10-17 17:50:07.015389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.015418] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.015784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.015815] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.016146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.016182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.016522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.016551] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.016917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.016947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.017381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.017410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.017774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.017803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.018155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.018183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.018514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.018541] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.018883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.018912] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.019264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.019292] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.019655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.019682] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.020043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.020071] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.020387] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.020414] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.020815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.020844] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.021200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.021227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.021594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.021623] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.022019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.022049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.022409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.022436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.022857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.022887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.023270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.023297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.023577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.023604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.023953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.023983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.024285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.024312] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.024682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.024722] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.025059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.025087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.025450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.025477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.025846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.025876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.026232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.026261] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.026597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.026625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.026854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.026887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.027257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.027286] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.027639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.027666] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.028026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.028055] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.028426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.028455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.028806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.028835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.029098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.029126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.029481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.029509] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.029860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.029890] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.030248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.398 [2024-10-17 17:50:07.030276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.398 qpair failed and we were unable to recover it. 00:28:59.398 [2024-10-17 17:50:07.030635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.030662] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.031025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.031055] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.031414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.031448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.031731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.031761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.032124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.032154] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.032524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.032552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.032816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.032845] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.033186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.033214] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.033659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.033687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.034065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.034093] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.034324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.034354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.034718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.034748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.035083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.035111] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.035475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.035503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.035867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.035896] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.036267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.036294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.036673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.036714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.036959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.036987] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.037365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.037393] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.037746] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.037775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.038155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.038183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.038588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.038615] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.038950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.038981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.039345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.039373] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.039750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.039779] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.040153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.040180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.040544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.040572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.040915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.040944] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.041304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.041332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.041597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.041626] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.041951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.041981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.042328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.042355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.042791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.042821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.043196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.043224] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.043641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.043669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.043995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.044024] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.044368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.044397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.044760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.044789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.045190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.045218] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.045618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.045647] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.399 qpair failed and we were unable to recover it. 00:28:59.399 [2024-10-17 17:50:07.045942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.399 [2024-10-17 17:50:07.045970] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.046309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.046337] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.046713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.046749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.047165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.047193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.047561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.047589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.047879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.047909] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.048279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.048306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.048670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.048707] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.049058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.049086] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.049451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.049478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.049841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.049870] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.050245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.050273] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.050631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.050659] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.051016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.051046] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.051409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.051436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.051800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.051829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.052195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.052224] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.052485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.052514] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.052884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.052913] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.053324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.053352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.053711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.053740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.054042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.054069] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.054296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.054328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.054569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.054599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.054933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.054963] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.055241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.055269] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.055510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.055539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.055993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.056024] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.056377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.056406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.056655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.056688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.057044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.057072] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.057445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.057473] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.057836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.057866] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.058113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.058144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.058495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.058524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.058874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.058904] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.059335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.059362] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.059721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.059753] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.060126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.060155] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.060517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.060545] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.400 [2024-10-17 17:50:07.060906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.400 [2024-10-17 17:50:07.060936] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.400 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.061292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.061320] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.061701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.061737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.062071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.062099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.062502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.062531] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.062915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.062945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.063292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.063320] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.063674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.063724] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.064102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.064130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.064498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.064527] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.064784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.064815] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.065156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.065184] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.065548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.065576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.065950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.065981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.066343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.066371] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.066743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.066772] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.067135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.067164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.067537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.067566] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.067942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.067972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.068273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.068301] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.068671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.068712] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.068964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.068996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.069364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.069392] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.069765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.069796] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.070171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.070199] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.070563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.070592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.070960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.070989] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.071351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.071379] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.071743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.071772] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.072156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.072184] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.072474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.072502] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.072867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.072897] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.073266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.073294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.073659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.073687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.074100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.074129] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.074389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.074420] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.074774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.074804] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.075177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.075205] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.075579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.075607] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.075861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.075895] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.076261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.401 [2024-10-17 17:50:07.076290] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.401 qpair failed and we were unable to recover it. 00:28:59.401 [2024-10-17 17:50:07.076658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.076687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.077102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.077138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.077269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.077301] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.077562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.077590] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.077944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.077973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.078245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.078273] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.078655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.078683] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.079048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.079077] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.079383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.079410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.079629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.079659] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.080049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.080080] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.080451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.080479] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.080731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.080764] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.081147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.081177] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.081524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.081552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.081809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.081839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.082200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.082228] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.082595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.082624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.082984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.083013] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.083387] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.083415] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.083672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.083714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.084095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.084124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.084486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.084514] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.084869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.084898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.085263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.085291] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.085660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.085690] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.086053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.086081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.086444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.086472] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.086906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.086937] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.087287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.087314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.087681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.087735] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.087999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.088027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.088252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.088279] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.088710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.088740] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.402 [2024-10-17 17:50:07.089099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.402 [2024-10-17 17:50:07.089127] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.402 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.089473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.089501] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.089845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.089874] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.090129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.090157] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.090506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.090534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.090909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.090938] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.091299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.091327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.091705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.091741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.091975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.092004] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.092359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.092386] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.092754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.092783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.093116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.093144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.093404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.093432] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.093839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.093869] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.094226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.094255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.094620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.094648] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.095017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.095046] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.095397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.095425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.095797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.095827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.096272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.096300] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.096666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.096705] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.097056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.097085] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.097453] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.097482] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.097866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.097897] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.098159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.098187] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.098649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.098676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.098931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.098959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.099318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.099346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.099597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.099626] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.099963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.099992] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.100358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.100385] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.100756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.100785] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.101146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.101174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.101427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.101455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.101861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.101891] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.102271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.102299] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.102659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.102687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.103067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.103097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.103449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.103477] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.103916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.103945] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.104303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.403 [2024-10-17 17:50:07.104332] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.403 qpair failed and we were unable to recover it. 00:28:59.403 [2024-10-17 17:50:07.104703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.104732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.105077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.105106] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.105476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.105505] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.105756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.105786] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.106133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.106160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.106536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.106564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.106922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.106958] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.107330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.107358] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.107720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.107749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.108100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.108128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.108439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.108467] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.108825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.108853] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.109215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.109243] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.109607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.109635] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.109996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.110026] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.110394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.110422] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.110775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.110805] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.111178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.111205] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.111546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.111574] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.111957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.111986] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.112252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.112281] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.112650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.112677] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.113051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.113079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.113511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.113540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.113903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.113934] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.114303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.114331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.114724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.114754] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.115115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.115143] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.115501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.115528] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.115900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.115929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.116298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.116325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.116687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.116726] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.117066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.117095] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.117463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.117492] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.117858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.117888] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.118253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.118282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.118640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.118668] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.119019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.119048] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.119407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.119435] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.119675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.119731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.404 [2024-10-17 17:50:07.120122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.404 [2024-10-17 17:50:07.120150] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.404 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.120528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.120556] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.120816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.120846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.121189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.121216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.121562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.121589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.121951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.121980] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.122345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.122373] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.122730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.122759] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.123148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.123177] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.123435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.123463] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.123739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.123768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.124116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.124144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.124506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.124533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.124892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.124922] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.125286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.125314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.125702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.125732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.126125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.126153] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.126510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.126539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.126926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.126955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.127308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.127336] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.127711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.127744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.128082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.128110] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.128476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.128504] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.128764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.128794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.129161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.129190] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.129611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.129641] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.130038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.130068] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.130376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.130405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.130777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.130806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.131173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.131201] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.131567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.131594] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.131893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.131924] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.132281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.132311] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.132702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.132738] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.133084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.133112] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.133475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.133503] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.133866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.133898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh: line 36: 233848 Killed "${NVMF_APP[@]}" "$@" 00:28:59.405 [2024-10-17 17:50:07.134294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.134325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.134703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.134733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 [2024-10-17 17:50:07.135015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.135045] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.405 qpair failed and we were unable to recover it. 00:28:59.405 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@48 -- # disconnect_init 10.0.0.2 00:28:59.405 [2024-10-17 17:50:07.135423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.405 [2024-10-17 17:50:07.135454] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:28:59.406 [2024-10-17 17:50:07.135946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.136048] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.406 [2024-10-17 17:50:07.136513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.136553] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.136957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.136990] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.137297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.137328] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.137597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.137627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.138019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.138051] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.138395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.138425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.138801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.138832] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.139205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.139233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.139612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.139642] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.140041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.140072] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.140453] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.140483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.140889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.140921] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.141275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.141304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.141686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.141731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.142089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.142119] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.142398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.142434] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.142735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.142766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.143141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.143170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.143537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.143567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.143728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.143759] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.144151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.144182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.144566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.144595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # nvmfpid=234876 00:28:59.406 [2024-10-17 17:50:07.144953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.144985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # waitforlisten 234876 00:28:59.406 [2024-10-17 17:50:07.145379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.145410] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@831 -- # '[' -z 234876 ']' 00:28:59.406 [2024-10-17 17:50:07.145773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.145804] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:59.406 [2024-10-17 17:50:07.146078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.146112] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:59.406 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:59.406 [2024-10-17 17:50:07.146419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.146450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:59.406 [2024-10-17 17:50:07.146832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.406 [2024-10-17 17:50:07.146864] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b9 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.406 0 with addr=10.0.0.2, port=4420 00:28:59.406 qpair failed and we were unable to recover it. 00:28:59.406 [2024-10-17 17:50:07.147275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.147308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.147678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.147719] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.148057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.148087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.148452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.148481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.148848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.148879] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.149278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.149309] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.149672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.149713] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.150070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.150100] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.150474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.150504] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.150871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.150913] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.151278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.151308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.151673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.151713] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.152055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.152090] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.152451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.152481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.152854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.152885] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.153252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.153282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.153679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.153720] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.154104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.154135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.154500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.154529] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.154909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.154941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.155322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.155353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.155601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.155630] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.155984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.156015] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.156422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.156453] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.156824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.156855] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.157096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.157126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.157415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.157445] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.157816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.157847] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.158228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.158258] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.158629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.158659] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.159065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.159096] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.159451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.159483] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.159774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.159807] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.160189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.160221] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.160578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.160612] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.160866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.160898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.161298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.161330] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.161702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.161732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.162079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.162107] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.162530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.162558] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.407 [2024-10-17 17:50:07.162934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.407 [2024-10-17 17:50:07.162966] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.407 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.163313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.163343] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.163714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.163744] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.164087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.164116] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.164496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.164524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.164867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.164896] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.165257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.165285] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.165677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.165717] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.166069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.166097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.166465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.166500] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.166737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.166767] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.167028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.167058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.167437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.167465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.167834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.167864] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.168230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.168260] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.168502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.168535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.168933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.168964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.169322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.169352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.169595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.169624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.170035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.170066] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.170287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.170319] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.170712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.170742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.171139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.171168] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.171540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.171569] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.171876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.171906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.172260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.172288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.172671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.172711] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.173123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.173152] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.173520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.173548] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.173893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.173923] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.174297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.174325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.174747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.174775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.175142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.175170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.175562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.175592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.175824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.175852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.176282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.176310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.176672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.176723] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.176969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.177001] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.177175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.177206] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.177553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.177582] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.177946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.408 [2024-10-17 17:50:07.177976] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.408 qpair failed and we were unable to recover it. 00:28:59.408 [2024-10-17 17:50:07.178354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.178382] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.178736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.178765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.179154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.179182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.179563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.179591] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.179962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.179992] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.180376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.180405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.180780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.180810] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.181195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.181223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.181476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.181511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.181887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.181917] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.182166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.182195] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.182561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.182589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.182941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.182970] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.183337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.183365] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.183733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.183762] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.184036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.184065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.184435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.184463] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.184838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.184868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.185277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.185305] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.185607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.185635] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.186020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.186051] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.186420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.186448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.186817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.186848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.187163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.187192] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.187417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.187449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.187820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.187849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.188227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.188255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.188508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.188536] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.188685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.188724] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.189135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.189163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.189527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.189555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.189911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.189940] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.190314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.190342] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.190722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.190752] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.191103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.191132] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.191533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.191561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.191799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.191829] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.192194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.192222] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.192582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.192610] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.192970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.192999] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.409 [2024-10-17 17:50:07.193364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.409 [2024-10-17 17:50:07.193393] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.409 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.193770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.193800] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.194223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.194251] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.194614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.194643] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.194984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.195014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.195381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.195409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.195771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.195801] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.196184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.196212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.196582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.196616] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.196962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.196991] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.197441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.197469] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.197850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.197879] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.198258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.198287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.198650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.198678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.199050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.199078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.199323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.199352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.199608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.199639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.200041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.200072] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.200438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.200467] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.200844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.200875] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.200955] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:28:59.410 [2024-10-17 17:50:07.201028] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:59.410 [2024-10-17 17:50:07.201240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.201279] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.201633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.201661] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.202047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.202079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.202440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.202470] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.202845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.202877] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.203238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.203269] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.203637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.203667] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.204054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.204086] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.204460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.204490] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.204857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.204889] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.205256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.205287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.205657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.205686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.205965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.205999] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.206272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.206303] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.206680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.206722] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.410 [2024-10-17 17:50:07.207085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.410 [2024-10-17 17:50:07.207114] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.410 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.207501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.207531] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.207921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.207952] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.208213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.208242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.208613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.208642] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.209019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.209051] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.209465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.209494] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.209949] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.209981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.210204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.210234] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.210562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.210592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.210940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.210971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.211341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.211372] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.211754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.211785] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.212162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.212191] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.212555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.212584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.212979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.213008] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.213265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.213299] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.213683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.213727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.214088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.214117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.214490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.214519] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.214907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.214937] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.215318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.215347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.215725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.215756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.216163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.216192] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.216420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.216452] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.216826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.216864] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.217215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.217243] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.217604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.217633] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.218001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.218033] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.218425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.218454] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.218714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.218745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.218993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.219022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.219318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.219346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.219687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.219729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.219966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.219996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.220390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.220419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.220686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.220733] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.220987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.221019] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.221291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.221320] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.221723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.221756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.222145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.222174] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.411 qpair failed and we were unable to recover it. 00:28:59.411 [2024-10-17 17:50:07.222426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.411 [2024-10-17 17:50:07.222454] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.222825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.222856] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.223237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.223266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.223505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.223538] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.223924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.223955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.224248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.224276] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.224647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.224676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.224849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.224878] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.225272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.225300] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.225678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.225723] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.226106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.226135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.226408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.226437] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.226776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.226806] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.227182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.227212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.227586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.227616] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.228022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.228053] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.228421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.228450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.228830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.228860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.229078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.229110] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.229372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.229406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.229778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.229808] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.230217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.230248] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.230616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.230646] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.230982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.231014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.231397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.231434] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.231798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.231830] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.232250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.232279] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.232655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.232685] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.232970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.232999] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.233371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.233402] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.233773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.233803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.234045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.234075] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.234456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.234485] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.234848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.234877] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.235147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.235177] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.235563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.235592] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.235997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.236027] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.236282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.236315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.236671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.236715] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.237089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.237120] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.412 [2024-10-17 17:50:07.237580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.412 [2024-10-17 17:50:07.237608] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.412 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.237847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.237876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.238244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.238271] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.238646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.238674] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.239049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.239078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.239446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.239473] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.239742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.239775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.240157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.240193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.240542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.240570] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.240885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.240914] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.241179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.241211] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.241553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.241583] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.241988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.242018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.242359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.242388] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.242609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.242639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.242928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.242960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.243299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.243329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.243715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.243747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.244120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.244153] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.244496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.244525] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.244943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.244973] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.245316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.245345] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.245686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.245726] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.246108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.246138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.246507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.246549] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.246891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.246922] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.247183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.247216] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.247589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.247621] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.247993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.248025] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.248406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.248436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.248652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.248684] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.249057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.249087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.249324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.249353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.249732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.249761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.250131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.250160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.250522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.250552] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.250781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.250813] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.251178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.251208] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.251587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.251617] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.251963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.251994] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.252355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.252384] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.413 qpair failed and we were unable to recover it. 00:28:59.413 [2024-10-17 17:50:07.252759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.413 [2024-10-17 17:50:07.252792] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.253178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.253207] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.253568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.253599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.253849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.253881] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.254251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.254280] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.254641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.254669] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.255047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.255078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.255424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.255453] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.255811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.255842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.256215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.256244] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.256592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.256622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.256997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.257028] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.257265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.257295] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.257681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.257721] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.258090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.258119] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.258496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.258525] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.258790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.258819] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.259057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.259090] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.259429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.259458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.259818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.259849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.260122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.260151] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.260527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.260557] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.260929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.260959] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.261324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.261359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.261620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.261650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.262083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.262115] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.262477] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.262506] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.262874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.262905] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.263293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.263323] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.263714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.263746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.264038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.264067] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.264412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.264441] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.264715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.264745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.265111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.265140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.265378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.265406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.265764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.265794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.266195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.266223] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.266596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.266625] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.266878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.266911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.267275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.414 [2024-10-17 17:50:07.267304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.414 qpair failed and we were unable to recover it. 00:28:59.414 [2024-10-17 17:50:07.267573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.267602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.267979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.268009] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.268383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.268412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.268670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.268712] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.269136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.269164] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.269519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.269548] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.269888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.269918] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.270303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.270331] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.270717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.270747] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.271137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.271165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.271532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.271561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.271951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.271981] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.272341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.272369] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.272448] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:28:59.415 [2024-10-17 17:50:07.272770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.272800] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.273065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.273094] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.273355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.273387] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.273713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.273743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.274104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.274134] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.274505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.274532] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.274919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.274949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.275334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.275363] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.275650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.275678] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.275975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.276004] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.276398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.276434] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.276816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.276846] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.277207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.277237] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.277479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.277507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.277758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.277791] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.278169] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.278198] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.278630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.278658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.279040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.279070] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.279434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.279463] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.279729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.279760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.280026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.280056] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.280420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.280448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.415 [2024-10-17 17:50:07.280720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.415 [2024-10-17 17:50:07.280750] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.415 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.281116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.281144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.281512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.281541] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.281904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.281935] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.282306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.282336] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.282710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.282741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.283099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.283128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.283499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.283527] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.283899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.283929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.284310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.284337] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.284713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.284743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.285045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.285074] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.285348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.285376] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.285758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.285789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.286145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.286175] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.286539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.286567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.286988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.287018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.287426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.287455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.287826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.287857] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.288205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.288233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.288599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.288628] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.288982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.289011] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.289374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.289402] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.289765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.289795] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.290165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.290193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.290557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.290585] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.290982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.291012] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.291377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.291405] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.291671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.291723] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.292083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.292112] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.292476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.292504] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.292868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.292897] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.293255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.293283] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.293645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.293673] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.293949] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.293978] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.294325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.294353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.294729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.294759] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.295114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.295142] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.295511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.295539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.295896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.295926] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.296152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.416 [2024-10-17 17:50:07.296180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.416 qpair failed and we were unable to recover it. 00:28:59.416 [2024-10-17 17:50:07.296534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.296562] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.296933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.296964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.297328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.297356] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.297718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.297748] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.298096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.298125] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.298400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.298428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.298787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.298817] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.299160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.299188] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.299528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.299556] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.299959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.299989] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.300340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.300369] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.300683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.300722] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.301086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.301114] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.301478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.301506] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.301866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.301896] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.302272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.302301] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.302667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.302709] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.302977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.303010] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.303242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.303274] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.303630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.303659] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.303978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.304010] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.304319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.304347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.304790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.304821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.417 [2024-10-17 17:50:07.305182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.417 [2024-10-17 17:50:07.305210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.417 qpair failed and we were unable to recover it. 00:28:59.687 [2024-10-17 17:50:07.305575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.687 [2024-10-17 17:50:07.305605] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.305991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.306021] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.306383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.306411] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.306815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.306852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.307182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.307210] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.307575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.307604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.307985] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.308014] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.308380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.308409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.308649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.308677] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.309043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.309071] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.309453] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.309482] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.309868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.309899] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.310273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.310302] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.310658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.310687] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.310961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.310990] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.311336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.311364] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.311592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.311620] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.311989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.312019] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.312380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.312408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.312883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.312913] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.313272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.313300] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.313687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.313729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.314065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.314094] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.314464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.314492] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.314869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.314901] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.315322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.315352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.315673] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:28:59.688 [2024-10-17 17:50:07.315724] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:28:59.688 [2024-10-17 17:50:07.315730] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:28:59.688 [2024-10-17 17:50:07.315736] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:28:59.688 [2024-10-17 17:50:07.315727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.315742] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:28:59.688 [2024-10-17 17:50:07.315758] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.316117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.316144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.316517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.316545] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.316928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.316957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.317339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.317367] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.317730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.317759] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 [2024-10-17 17:50:07.317628] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.317770] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:28:59.688 [2024-10-17 17:50:07.318164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.318192] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.318159] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:28:59.688 [2024-10-17 17:50:07.318160] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:28:59.688 [2024-10-17 17:50:07.318567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.318595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.318960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.318990] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.319369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.319397] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.319622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.688 [2024-10-17 17:50:07.319650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.688 qpair failed and we were unable to recover it. 00:28:59.688 [2024-10-17 17:50:07.319883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.319913] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.320192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.320220] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.320575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.320602] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.321068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.321099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.321364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.321392] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.321758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.321788] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.322158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.322186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.322467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.322495] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.322841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.322872] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.323233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.323261] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.323510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.323539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.323889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.323918] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.324286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.324314] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.324576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.324604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.324851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.324880] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.325226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.325255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.325617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.325653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.326054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.326084] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.326325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.326354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.326731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.326762] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.327131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.327160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.327527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.327555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.327824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.327854] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.328234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.328262] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.328622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.328650] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.329010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.329040] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.329160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.329187] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.329620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.329648] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.329912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.329941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.330161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.330189] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.330561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.330589] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.330821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.330852] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.331106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.331134] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.331402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.331430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.331672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.331708] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.331956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.331985] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.332362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.332391] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.332760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.332789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.333141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.333169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.333513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.333541] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.689 [2024-10-17 17:50:07.333642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.689 [2024-10-17 17:50:07.333671] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.689 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.333926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.333956] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.334313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.334341] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.334576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.334604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.334838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.334868] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.335116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.335144] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.335491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.335521] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.335767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.335796] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.336142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.336170] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.336548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.336576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.336928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.336957] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.337322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.337350] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.337719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.337749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.337849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.337876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.338213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.338241] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.338484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.338513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.338721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.338757] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.339032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.339060] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.339282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.339311] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.339703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.339734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.339974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.340006] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.340377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.340407] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.340660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.340689] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.341059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.341088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.341342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.341372] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.341730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.341760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.342098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.342126] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.342399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.342431] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.342772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.342803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.343061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.343090] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.343329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.343359] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.343609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.343639] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.343990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.344021] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.344385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.344414] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.344792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.344821] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.345070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.345102] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.345429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.345457] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.345660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.345689] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.345962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.345995] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.346236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.346265] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.346642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.690 [2024-10-17 17:50:07.346670] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.690 qpair failed and we were unable to recover it. 00:28:59.690 [2024-10-17 17:50:07.346883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.346913] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.347372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.347401] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.347742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.347773] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.348135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.348163] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.348529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.348557] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.348819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.348849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.349071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.349099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.349321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.349353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.349724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.349755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.350118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.350148] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.350513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.350542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.350814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.350845] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.351212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.351242] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.351367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.351398] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.351507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.351535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa898000b90 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 [2024-10-17 17:50:07.352366] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:28:59.691 [2024-10-17 17:50:07.352863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.352912] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1e690d0 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.353298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.353330] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1e690d0 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 [2024-10-17 17:50:07.353438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.691 [2024-10-17 17:50:07.353466] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1e690d0 with addr=10.0.0.2, port=4420 00:28:59.691 qpair failed and we were unable to recover it. 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Write completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 Read completed with error (sct=0, sc=8) 00:28:59.691 starting I/O failed 00:28:59.691 [2024-10-17 17:50:07.354266] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:28:59.691 [2024-10-17 17:50:07.354720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.354789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.355167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.355265] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.355721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.355761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.356128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.356227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.356622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.356660] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.357270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.357368] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.357932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.358030] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.358312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.358349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.358727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.358761] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.359003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.359032] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.359436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.359465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.359803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.359835] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.360205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.360234] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.360469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.360498] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.360751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.360784] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.361158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.361186] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.361327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.361355] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.361713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.361743] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.362109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.362138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.362379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.362407] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.362752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.362783] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.362883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.362909] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.363254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.363283] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.363529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.363560] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.363939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.363969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.364288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.364316] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.364667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.364706] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.365141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.365169] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.365374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.365402] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.365762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.365791] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.366157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.366185] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.366566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.366594] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.366726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.366756] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.367023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.367052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.692 qpair failed and we were unable to recover it. 00:28:59.692 [2024-10-17 17:50:07.367424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.692 [2024-10-17 17:50:07.367452] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.367668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.367710] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.368077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.368105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.368496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.368530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.368747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.368778] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.369001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.369029] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.369248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.369278] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.369508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.369544] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.369784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.369814] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.370197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.370225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.370532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.370559] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.370917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.370947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.371321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.371349] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.371716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.371746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.372096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.372124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.372497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.372526] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.372765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.372794] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.373148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.373177] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.373408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.373436] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.373800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.373830] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.373934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.373962] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.374358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.374386] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.374627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.374655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.375023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.375052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.375402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.375430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.375801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.375832] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.376038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.376066] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.376388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.376416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.376634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.376662] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.376900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.376929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.377261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.377289] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.377512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.377539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.377646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.377671] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.377927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.377955] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.378356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.378383] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.378611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.378640] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.379022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.379052] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.379427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.379455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.379724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.379755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.379994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.693 [2024-10-17 17:50:07.380022] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.693 qpair failed and we were unable to recover it. 00:28:59.693 [2024-10-17 17:50:07.380288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.380316] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.380653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.380681] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.380894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.380922] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.381282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.381317] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.381527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.381556] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.381777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.381805] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.382156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.382184] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.382401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.382429] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.382767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.382796] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.382939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.382966] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.383059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.383088] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.383440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.383467] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.383847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.383877] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.384259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.384287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.384549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.384577] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.384807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.384837] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.385103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.385130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.385484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.385512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.385861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.385891] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.386293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.386321] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.386540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.386573] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.386831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.386860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.387221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.387249] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.387502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.387533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.387933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.387964] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.388182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.388214] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.388570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.388598] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.388820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.388849] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.389081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.389108] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.389481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.389509] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.389769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.389800] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.390015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.390043] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.390402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.390430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.390793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.390822] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.391212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.391240] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.391501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.391530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.391905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.391935] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.392113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.392141] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.392514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.392542] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.392648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.392679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.392919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.694 [2024-10-17 17:50:07.392948] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.694 qpair failed and we were unable to recover it. 00:28:59.694 [2024-10-17 17:50:07.393152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.393182] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.393506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.393535] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.393747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.393782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.394025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.394053] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.394307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.394334] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.394689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.394727] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.395075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.395103] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.395343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.395371] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.395735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.395765] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.396009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.396041] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.396462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.396490] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.396865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.396895] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.397136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.397167] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.397539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.397567] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.397920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.397949] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.398211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.398239] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.398648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.398676] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.399056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.399085] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.399457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.399485] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.399580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.399610] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa8a4000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Write completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 Read completed with error (sct=0, sc=8) 00:28:59.695 starting I/O failed 00:28:59.695 [2024-10-17 17:50:07.400380] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.695 [2024-10-17 17:50:07.401026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.401136] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.401572] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.401608] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.401980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.402087] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.402526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.402563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.402934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.402966] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.403214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.403243] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.403479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.403508] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.403900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.403929] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.404208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.404236] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.695 qpair failed and we were unable to recover it. 00:28:59.695 [2024-10-17 17:50:07.404629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.695 [2024-10-17 17:50:07.404657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.404800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.404831] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.405072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.405101] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.405555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.405583] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.405988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.406018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.406387] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.406416] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.406799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.406828] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.407220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.407249] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.407479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.407507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.407909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.407941] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.408282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.408310] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.408428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.408455] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.408711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.408741] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.409101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.409128] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.409495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.409524] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.409957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.409986] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.410352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.410380] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.410626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.410658] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.411049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.411081] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.411404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.411433] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.411775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.411807] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.412029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.412058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.412421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.412448] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.412810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.412839] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.413209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.413238] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.413481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.413509] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.413859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.413889] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.414258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.414287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:59.696 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # return 0 00:28:59.696 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:28:59.696 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:28:59.696 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.696 [2024-10-17 17:50:07.417396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.417458] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.417847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.417882] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.418259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.418288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.418654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.418700] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.419003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.419032] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.419390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.419419] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.696 qpair failed and we were unable to recover it. 00:28:59.696 [2024-10-17 17:50:07.419864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.696 [2024-10-17 17:50:07.419894] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.420263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.420292] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.420657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.420686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.421070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.421099] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.421444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.421474] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.421841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.421873] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.422205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.422233] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.422418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.422447] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.422677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.422714] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.423091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.423120] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.423493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.423521] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.423937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.423969] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.424324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.424352] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.424703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.424734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.424954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.424983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.425206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.425234] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.425622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.425652] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.426034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.426065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.426447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.426475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.426856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.426887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.427246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.427275] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.427473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.427505] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.427898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.427930] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.428147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.428176] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.428536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.428564] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.428934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.428965] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.429175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.429204] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.429556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.429584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.429857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.429889] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.430005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.430037] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.430265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.430295] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.430543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.430571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.430658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.430686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.431019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.431048] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.431381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.431409] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.431774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.431803] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.432130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.432160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.432537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.432571] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.432944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.432976] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.433334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.433362] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.433599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.697 [2024-10-17 17:50:07.433629] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.697 qpair failed and we were unable to recover it. 00:28:59.697 [2024-10-17 17:50:07.433883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.433911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.434269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.434297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.434512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.434539] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.434908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.434938] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.435260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.435287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.435662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.435718] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.435959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.435988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.436226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.436255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.436503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.436531] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.436762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.436796] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.437043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.437073] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.437319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.437347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.437715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.437745] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.437983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.438012] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.438367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.438396] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.438599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.438627] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.438991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.439020] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.439256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.439287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.439637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.439666] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.440024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.440053] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.440489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.440517] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.440768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.440799] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.441010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.441038] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.441398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.441425] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.441791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.441822] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.442030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.442058] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.442419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.442449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.442796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.442826] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.443090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.443118] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.443547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.443576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.443795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.443827] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.444198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.444227] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.444449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.444478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.444609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.444636] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.445018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.445049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.445257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.445287] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.445733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.445770] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.446095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.446123] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.446485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.446513] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.446871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.446900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.698 [2024-10-17 17:50:07.447121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.698 [2024-10-17 17:50:07.447148] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.698 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.447502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.447530] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.447759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.447789] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.448152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.448180] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.448539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.448566] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.448923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.448953] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.449319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.449346] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.449575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.449604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.450020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.450050] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.450302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.450330] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.450679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.450719] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.451063] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.451093] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.451449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.451478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.451741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.451772] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.452030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.452059] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.452401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.452430] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.452786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.452816] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.453037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.453065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.453421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.453449] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.453869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.453900] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.454144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.454179] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.454514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.454543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.454709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.454746] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.455103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.455132] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.455495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.455523] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.455889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.455920] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.456249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.456277] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.456637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.456665] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.457036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.457065] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.457372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.457401] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.457625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.457653] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.458065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.458097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.458337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.458365] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:28:59.699 [2024-10-17 17:50:07.458568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.458599] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.458957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:28:59.699 [2024-10-17 17:50:07.458988] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.699 [2024-10-17 17:50:07.459357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.459389] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.699 [2024-10-17 17:50:07.459547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.459576] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.459832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.459862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.460198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.460225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.699 qpair failed and we were unable to recover it. 00:28:59.699 [2024-10-17 17:50:07.460593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.699 [2024-10-17 17:50:07.460621] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.460988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.461018] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.461275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.461304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.461660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.461688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.462068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.462096] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.462461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.462488] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.462748] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.462782] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.463154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.463183] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.463514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.463541] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.463903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.463932] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.464290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.464319] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.464675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.464713] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.464924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.464951] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.465174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.465207] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.465449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.465481] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.465853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.465884] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.466254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.466282] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.466614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.466642] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.466965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.466999] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.467371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.467400] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.467620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.467647] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.467984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.468013] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.468385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.468414] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.468649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.468679] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.468918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.468947] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.469297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.469325] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.469678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.469732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.470096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.470124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.470483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.470511] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.470785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.470815] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.471052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.471080] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.471459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.471487] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.471747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.471775] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.472134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.472162] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.472516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.472544] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.472948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.472983] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.473187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.700 [2024-10-17 17:50:07.473215] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.700 qpair failed and we were unable to recover it. 00:28:59.700 [2024-10-17 17:50:07.473423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.473451] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.473699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.473728] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.474107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.474135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.474382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.474414] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.474628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.474655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.474924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.474954] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.475320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.475347] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.475567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.475595] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.475858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.475892] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.476217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.476245] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.476473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.476505] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.476851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.476880] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.477244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.477272] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.477480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.477508] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.477737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.477766] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.478029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.478057] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.478395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.478422] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.478641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.478668] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.478891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.478920] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.479278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.479306] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.479660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.479688] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.479942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.479971] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.480174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.480202] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.480556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.480584] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.480973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.481003] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.481404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.481432] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.481888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.481917] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.482148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.482179] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.482533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.482561] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.482914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.482943] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.483302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.483329] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.483552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.483581] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.483949] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.483977] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.484324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.484351] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.484721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.484750] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.485089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.485117] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.485474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.485502] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.485721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.485749] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.486027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.486059] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.486444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.486472] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.701 qpair failed and we were unable to recover it. 00:28:59.701 [2024-10-17 17:50:07.486747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.701 [2024-10-17 17:50:07.486776] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.487036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.487064] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.487288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.487316] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.487740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.487768] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.487965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.487993] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.488265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.488297] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.488644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.488672] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.488937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.488965] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.489311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.489339] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.489735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.489767] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.490132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.490160] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.490282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.490308] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.490702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.490732] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.491081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.491109] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.491467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.491496] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.491730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.491763] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.492110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.492138] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.492400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.492428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.492785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.492814] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.493184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.493212] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.493419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.493447] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.493659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.493686] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.493941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.493972] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.494342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.494370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.494725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.494755] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.495016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.495045] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 Malloc0 00:28:59.702 [2024-10-17 17:50:07.495398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.495426] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.495789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.495819] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.702 [2024-10-17 17:50:07.496047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.496079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:28:59.702 [2024-10-17 17:50:07.496447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.496475] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.702 [2024-10-17 17:50:07.496708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.496737] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.702 [2024-10-17 17:50:07.497095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.497124] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.497437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.497465] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.497730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.497760] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.497979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.498010] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.498246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.498275] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.498627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.498655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.498881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.498912] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.499163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.702 [2024-10-17 17:50:07.499193] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.702 qpair failed and we were unable to recover it. 00:28:59.702 [2024-10-17 17:50:07.499428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.499457] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.499817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.499848] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.500213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.500241] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.500479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.500507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.500833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.500862] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.501197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.501225] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.501594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.501622] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.501977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.502007] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.502362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.502389] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.502522] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:59.703 [2024-10-17 17:50:07.502604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.502632] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.502964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.502996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.503354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.503383] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.503747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.503777] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.504016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.504044] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.504404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.504431] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.504672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.504728] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.505107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.505135] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.505376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.505408] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.505805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.505834] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.506193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.506222] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.506474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.506506] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.506852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.506881] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.507227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.507255] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.507629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.507657] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.508049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.508079] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.508299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.508327] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.508704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.508734] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.509057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.509085] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.509191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.509217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.509545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.509572] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.509931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.509960] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.510400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.510428] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.510847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.510876] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.511084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.511112] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 [2024-10-17 17:50:07.511323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.511353] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.703 [2024-10-17 17:50:07.511623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.511655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:28:59.703 [2024-10-17 17:50:07.512000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.512036] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.703 [2024-10-17 17:50:07.512276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.512304] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.703 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.703 [2024-10-17 17:50:07.512515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.703 [2024-10-17 17:50:07.512543] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.703 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.512831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.512860] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.513113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.513140] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.513505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.513533] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.513742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.513772] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.514035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.514064] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.514422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.514450] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.514821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.514850] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.515190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.515217] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.515484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.515512] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.515858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.515887] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.516240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.516268] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.516479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.516507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.516856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.516885] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.517131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.517162] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.517507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.517534] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.517898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.517928] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.518384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.518412] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.518607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.518635] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.518867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.518898] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.519238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.519266] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.519624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.519651] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.519882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.519911] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.520143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.520173] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.520284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.520315] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.520596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.520624] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.520960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.520989] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.521342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.521370] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.521745] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.521776] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.522137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.522165] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.522535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.522563] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.522967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.522996] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.523378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.523406] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.704 [2024-10-17 17:50:07.523764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.523793] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:28:59.704 [2024-10-17 17:50:07.524021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.524049] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.704 [2024-10-17 17:50:07.524412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.524440] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.704 [2024-10-17 17:50:07.524782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.524811] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.525168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.525196] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.525569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.704 [2024-10-17 17:50:07.525597] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.704 qpair failed and we were unable to recover it. 00:28:59.704 [2024-10-17 17:50:07.525951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.525980] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.526235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.526262] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.526509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.526540] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.526789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.526818] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.527261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.527288] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.527700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.527729] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.528078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.528105] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.528441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.528468] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.528830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.528859] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.529221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.529248] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.529617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.529645] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.530014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.530043] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.530458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.530485] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.530826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.530856] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.530965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.530991] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.531451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.531478] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.531702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.531731] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.531843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.531873] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.531967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.532000] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.532322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.532350] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.532583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.532611] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.532815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.532842] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.533089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.533118] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.533473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.533507] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.533854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.533883] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.534120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.534147] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.534497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.534525] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.534808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.534838] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.535069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.535097] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 [2024-10-17 17:50:07.535326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.535354] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.705 [2024-10-17 17:50:07.535712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.535742] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:59.705 [2024-10-17 17:50:07.536091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.536119] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.705 [2024-10-17 17:50:07.536492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.536520] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.705 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.705 [2024-10-17 17:50:07.536625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.705 [2024-10-17 17:50:07.536655] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.705 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.537049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.537078] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.537528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.537556] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.537775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.537807] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.538180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.538208] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.538577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.538604] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.538876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.538906] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.539266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.539294] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.539527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.539555] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.539922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.539950] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.540311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.540339] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.540730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.540758] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.541101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.541130] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.541476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.541505] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.541791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.541820] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.542170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.542198] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.542574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:28:59.706 [2024-10-17 17:50:07.542601] nvme_tcp.c:2399:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fa89c000b90 with addr=10.0.0.2, port=4420 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.542770] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:59.706 [2024-10-17 17:50:07.543842] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.706 [2024-10-17 17:50:07.543977] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.706 [2024-10-17 17:50:07.544023] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.706 [2024-10-17 17:50:07.544045] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.706 [2024-10-17 17:50:07.544066] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.706 [2024-10-17 17:50:07.544117] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.706 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:28:59.706 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.706 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:28:59.706 [2024-10-17 17:50:07.553446] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.706 [2024-10-17 17:50:07.553569] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.706 [2024-10-17 17:50:07.553608] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.706 [2024-10-17 17:50:07.553630] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.706 [2024-10-17 17:50:07.553658] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.706 [2024-10-17 17:50:07.553721] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.706 17:50:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@50 -- # wait 234201 00:28:59.706 [2024-10-17 17:50:07.563418] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.706 [2024-10-17 17:50:07.563528] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.706 [2024-10-17 17:50:07.563554] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.706 [2024-10-17 17:50:07.563569] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.706 [2024-10-17 17:50:07.563582] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.706 [2024-10-17 17:50:07.563616] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.573317] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.706 [2024-10-17 17:50:07.573378] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.706 [2024-10-17 17:50:07.573397] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.706 [2024-10-17 17:50:07.573407] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.706 [2024-10-17 17:50:07.573416] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.706 [2024-10-17 17:50:07.573435] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.583340] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.706 [2024-10-17 17:50:07.583398] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.706 [2024-10-17 17:50:07.583412] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.706 [2024-10-17 17:50:07.583419] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.706 [2024-10-17 17:50:07.583425] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.706 [2024-10-17 17:50:07.583440] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.706 [2024-10-17 17:50:07.593356] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.706 [2024-10-17 17:50:07.593402] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.706 [2024-10-17 17:50:07.593416] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.706 [2024-10-17 17:50:07.593423] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.706 [2024-10-17 17:50:07.593429] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.706 [2024-10-17 17:50:07.593444] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.706 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.603382] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.603429] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.603443] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.603450] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.603457] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.603471] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.613362] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.613411] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.613428] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.613435] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.613441] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.613456] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.623344] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.623409] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.623423] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.623430] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.623436] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.623450] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.633468] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.633516] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.633530] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.633537] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.633543] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.633557] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.643481] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.643528] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.643541] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.643548] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.643554] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.643568] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.653461] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.653507] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.653520] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.653526] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.653536] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.653550] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.663411] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.663460] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.663475] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.663482] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.663489] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.663503] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.673576] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.673640] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.673654] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.673661] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.673667] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.673681] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.683563] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.683617] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.683632] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.683639] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.683646] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.683660] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.693555] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.693600] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.693613] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.693620] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.693627] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.693641] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.703624] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.703678] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.703695] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.703702] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.703709] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.703723] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.968 [2024-10-17 17:50:07.713523] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.968 [2024-10-17 17:50:07.713572] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.968 [2024-10-17 17:50:07.713587] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.968 [2024-10-17 17:50:07.713594] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.968 [2024-10-17 17:50:07.713600] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.968 [2024-10-17 17:50:07.713614] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.968 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.723653] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.723704] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.723718] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.723725] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.723732] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.723746] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.733663] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.733757] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.733795] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.733803] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.733810] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.733834] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.743742] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.743794] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.743809] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.743816] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.743827] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.743842] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.753718] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.753769] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.753783] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.753790] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.753796] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.753810] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.763789] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.763840] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.763853] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.763860] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.763867] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.763881] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.773654] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.773703] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.773717] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.773723] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.773730] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.773744] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.783833] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.783885] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.783899] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.783905] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.783912] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.783926] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.793876] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.793926] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.793939] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.793946] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.793952] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.793966] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.803895] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.803944] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.803957] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.803964] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.803970] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.803984] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.813869] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.813912] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.813926] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.813932] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.813939] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.813953] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.823833] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.823892] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.823907] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.823914] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.823921] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.823938] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.833930] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.833975] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.833988] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.833999] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.834005] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.834020] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.843880] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.843939] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.843953] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.843960] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.843966] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.843981] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.853976] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.969 [2024-10-17 17:50:07.854020] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.969 [2024-10-17 17:50:07.854033] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.969 [2024-10-17 17:50:07.854040] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.969 [2024-10-17 17:50:07.854046] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.969 [2024-10-17 17:50:07.854060] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.969 qpair failed and we were unable to recover it. 00:28:59.969 [2024-10-17 17:50:07.864057] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.970 [2024-10-17 17:50:07.864113] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.970 [2024-10-17 17:50:07.864126] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.970 [2024-10-17 17:50:07.864133] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.970 [2024-10-17 17:50:07.864139] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.970 [2024-10-17 17:50:07.864153] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.970 qpair failed and we were unable to recover it. 00:28:59.970 [2024-10-17 17:50:07.873927] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.970 [2024-10-17 17:50:07.873974] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.970 [2024-10-17 17:50:07.873987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.970 [2024-10-17 17:50:07.873994] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.970 [2024-10-17 17:50:07.874000] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.970 [2024-10-17 17:50:07.874014] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.970 qpair failed and we were unable to recover it. 00:28:59.970 [2024-10-17 17:50:07.883975] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:28:59.970 [2024-10-17 17:50:07.884044] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:28:59.970 [2024-10-17 17:50:07.884057] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:28:59.970 [2024-10-17 17:50:07.884064] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:28:59.970 [2024-10-17 17:50:07.884070] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:28:59.970 [2024-10-17 17:50:07.884084] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:28:59.970 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.894092] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.894139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.894152] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.894159] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.894165] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.894179] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.904173] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.904253] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.904266] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.904273] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.904279] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.904294] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.914181] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.914226] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.914240] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.914246] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.914253] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.914267] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.924193] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.924246] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.924258] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.924268] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.924275] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.924289] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.934205] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.934249] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.934262] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.934269] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.934275] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.934289] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.944280] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.944336] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.944349] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.944356] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.944362] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.944376] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.954284] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.954335] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.954348] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.954356] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.954362] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.954375] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.964310] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.964384] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.964397] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.964404] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.964411] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.964424] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.974362] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.974410] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.974423] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.974430] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.974436] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.974450] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.984428] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.984485] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.984498] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.232 [2024-10-17 17:50:07.984505] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.232 [2024-10-17 17:50:07.984511] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.232 [2024-10-17 17:50:07.984525] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.232 qpair failed and we were unable to recover it. 00:29:00.232 [2024-10-17 17:50:07.994406] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.232 [2024-10-17 17:50:07.994458] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.232 [2024-10-17 17:50:07.994471] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:07.994478] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:07.994484] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:07.994498] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.004421] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.004467] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.004480] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.004488] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.004494] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.004508] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.014438] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.014484] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.014501] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.014508] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.014514] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.014528] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.024495] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.024552] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.024565] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.024572] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.024579] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.024592] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.034544] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.034611] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.034623] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.034630] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.034637] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.034651] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.044524] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.044570] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.044583] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.044590] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.044597] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.044612] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.054538] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.054591] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.054604] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.054611] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.054618] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.054635] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.064600] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.064653] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.064666] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.064673] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.064680] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.064697] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.074621] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.074671] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.074684] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.074696] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.074703] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.074717] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.084671] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.084746] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.084759] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.084766] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.084773] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.084787] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.094509] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.094556] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.094571] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.094579] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.094586] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.094601] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.104733] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.104799] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.104815] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.104823] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.104829] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.104844] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.114714] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.114764] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.114778] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.114785] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.114791] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.114805] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.124726] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.124777] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.124790] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.124797] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.233 [2024-10-17 17:50:08.124803] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.233 [2024-10-17 17:50:08.124818] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.233 qpair failed and we were unable to recover it. 00:29:00.233 [2024-10-17 17:50:08.134727] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.233 [2024-10-17 17:50:08.134783] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.233 [2024-10-17 17:50:08.134796] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.233 [2024-10-17 17:50:08.134803] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.234 [2024-10-17 17:50:08.134809] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.234 [2024-10-17 17:50:08.134823] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.234 qpair failed and we were unable to recover it. 00:29:00.234 [2024-10-17 17:50:08.144685] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.234 [2024-10-17 17:50:08.144754] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.234 [2024-10-17 17:50:08.144769] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.234 [2024-10-17 17:50:08.144776] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.234 [2024-10-17 17:50:08.144783] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.234 [2024-10-17 17:50:08.144812] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.234 qpair failed and we were unable to recover it. 00:29:00.495 [2024-10-17 17:50:08.154833] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.495 [2024-10-17 17:50:08.154883] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.495 [2024-10-17 17:50:08.154897] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.495 [2024-10-17 17:50:08.154904] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.495 [2024-10-17 17:50:08.154910] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.495 [2024-10-17 17:50:08.154925] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.495 qpair failed and we were unable to recover it. 00:29:00.495 [2024-10-17 17:50:08.164734] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.495 [2024-10-17 17:50:08.164791] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.495 [2024-10-17 17:50:08.164804] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.495 [2024-10-17 17:50:08.164811] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.495 [2024-10-17 17:50:08.164818] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.495 [2024-10-17 17:50:08.164837] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.495 qpair failed and we were unable to recover it. 00:29:00.495 [2024-10-17 17:50:08.174838] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.495 [2024-10-17 17:50:08.174886] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.495 [2024-10-17 17:50:08.174900] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.495 [2024-10-17 17:50:08.174907] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.495 [2024-10-17 17:50:08.174913] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.174929] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.184831] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.184890] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.184904] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.184911] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.184917] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.184931] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.194952] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.195002] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.195015] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.195022] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.195028] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.195042] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.204951] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.205001] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.205014] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.205021] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.205027] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.205041] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.214978] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.215027] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.215040] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.215047] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.215053] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.215067] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.225041] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.225139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.225152] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.225159] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.225165] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.225178] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.235056] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.235104] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.235117] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.235124] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.235133] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.235147] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.245060] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.245110] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.245123] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.245129] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.245136] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.245150] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.255065] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.255113] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.255126] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.255133] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.255139] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.255152] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.265046] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.265102] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.265114] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.265122] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.265128] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.265142] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.275161] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.275210] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.275223] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.275230] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.275236] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.275250] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.285177] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.285241] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.285255] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.285262] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.285268] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.285282] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.295174] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.295222] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.295235] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.295242] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.295248] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.295262] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.305267] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.305322] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.305334] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.305341] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.305348] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.305361] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.496 [2024-10-17 17:50:08.315271] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.496 [2024-10-17 17:50:08.315318] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.496 [2024-10-17 17:50:08.315332] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.496 [2024-10-17 17:50:08.315339] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.496 [2024-10-17 17:50:08.315345] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.496 [2024-10-17 17:50:08.315359] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.496 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.325435] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.325506] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.325519] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.325529] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.325535] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.325549] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.335327] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.335376] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.335390] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.335396] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.335403] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.335416] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.345419] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.345473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.345485] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.345492] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.345499] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.345512] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.355421] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.355471] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.355484] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.355491] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.355497] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.355511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.365398] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.365448] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.365462] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.365469] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.365475] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.365490] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.375416] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.375504] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.375518] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.375525] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.375531] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.375545] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.385475] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.385529] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.385542] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.385549] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.385555] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.385569] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.395452] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.395503] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.395516] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.395523] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.395529] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.395543] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.497 [2024-10-17 17:50:08.405493] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.497 [2024-10-17 17:50:08.405541] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.497 [2024-10-17 17:50:08.405554] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.497 [2024-10-17 17:50:08.405561] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.497 [2024-10-17 17:50:08.405567] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.497 [2024-10-17 17:50:08.405581] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.497 qpair failed and we were unable to recover it. 00:29:00.759 [2024-10-17 17:50:08.415513] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.759 [2024-10-17 17:50:08.415559] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.759 [2024-10-17 17:50:08.415572] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.759 [2024-10-17 17:50:08.415582] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.759 [2024-10-17 17:50:08.415589] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.759 [2024-10-17 17:50:08.415602] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.759 qpair failed and we were unable to recover it. 00:29:00.759 [2024-10-17 17:50:08.425576] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.425629] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.425642] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.425649] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.425656] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.425670] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.435601] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.435649] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.435662] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.435669] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.435675] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.435689] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.445606] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.445659] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.445672] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.445679] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.445685] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.445703] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.455663] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.455738] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.455751] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.455759] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.455765] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.455778] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.465556] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.465650] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.465663] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.465670] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.465676] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.465695] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.475720] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.475773] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.475786] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.475793] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.475800] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.475814] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.485619] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.485668] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.485681] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.485689] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.485699] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.485712] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.495773] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.495827] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.495840] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.495847] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.495853] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.495867] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.505807] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.505858] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.505874] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.505881] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.505888] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.505902] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.515702] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.515794] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.515808] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.515815] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.515821] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.515835] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.525830] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.525881] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.525894] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.525901] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.525907] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.525922] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.535828] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.535884] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.535897] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.535904] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.535911] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.535925] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.545914] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.545963] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.545976] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.545983] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.545989] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.546006] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.555934] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.555985] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.760 [2024-10-17 17:50:08.555998] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.760 [2024-10-17 17:50:08.556005] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.760 [2024-10-17 17:50:08.556011] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.760 [2024-10-17 17:50:08.556025] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.760 qpair failed and we were unable to recover it. 00:29:00.760 [2024-10-17 17:50:08.565942] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.760 [2024-10-17 17:50:08.566039] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.566052] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.566059] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.566065] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.566079] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.575945] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.575997] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.576010] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.576017] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.576024] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.576037] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.586018] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.586073] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.586086] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.586092] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.586099] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.586112] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.596025] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.596122] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.596138] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.596145] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.596152] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.596165] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.606018] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.606067] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.606080] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.606087] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.606093] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.606107] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.616057] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.616111] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.616124] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.616131] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.616138] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.616151] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.626127] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.626178] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.626191] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.626198] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.626204] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.626218] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.636017] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.636065] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.636079] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.636086] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.636092] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.636118] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.646173] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.646221] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.646234] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.646241] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.646247] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.646261] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.656162] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.656213] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.656226] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.656233] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.656240] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.656253] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.666217] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.666271] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.666284] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.666291] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.666297] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.666311] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:00.761 [2024-10-17 17:50:08.676235] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:00.761 [2024-10-17 17:50:08.676286] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:00.761 [2024-10-17 17:50:08.676299] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:00.761 [2024-10-17 17:50:08.676306] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:00.761 [2024-10-17 17:50:08.676313] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:00.761 [2024-10-17 17:50:08.676326] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:00.761 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.686132] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.686181] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.686197] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.686204] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.686210] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.023 [2024-10-17 17:50:08.686224] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.023 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.696266] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.696310] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.696323] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.696330] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.696336] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.023 [2024-10-17 17:50:08.696350] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.023 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.706343] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.706398] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.706411] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.706418] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.706424] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.023 [2024-10-17 17:50:08.706438] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.023 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.716362] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.716410] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.716423] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.716430] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.716436] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.023 [2024-10-17 17:50:08.716450] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.023 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.726291] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.726341] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.726354] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.726361] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.726371] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.023 [2024-10-17 17:50:08.726385] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.023 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.736331] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.736378] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.736392] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.736399] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.736405] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.023 [2024-10-17 17:50:08.736419] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.023 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.746440] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.746533] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.746546] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.746552] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.746559] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.023 [2024-10-17 17:50:08.746573] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.023 qpair failed and we were unable to recover it. 00:29:01.023 [2024-10-17 17:50:08.756464] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.023 [2024-10-17 17:50:08.756553] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.023 [2024-10-17 17:50:08.756566] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.023 [2024-10-17 17:50:08.756573] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.023 [2024-10-17 17:50:08.756579] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.756593] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.766479] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.766527] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.766540] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.766547] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.766553] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.766567] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.776483] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.776530] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.776543] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.776550] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.776556] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.776570] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.786530] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.786582] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.786595] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.786602] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.786608] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.786622] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.796562] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.796608] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.796622] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.796629] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.796635] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.796648] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.806602] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.806647] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.806660] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.806667] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.806673] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.806687] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.816634] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.816710] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.816724] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.816731] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.816741] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.816757] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.826659] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.826717] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.826730] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.826737] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.826744] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.826758] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.836672] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.836728] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.836741] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.836748] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.836754] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.836768] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.846693] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.846741] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.846754] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.846762] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.846768] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.846782] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.856657] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.856713] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.856726] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.856733] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.856739] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.856753] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.866627] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.866675] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.866688] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.866700] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.866706] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.866727] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.876737] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.876804] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.876818] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.876825] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.876831] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.876845] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.886677] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.886727] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.886740] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.886747] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.886753] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.886767] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.896786] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.024 [2024-10-17 17:50:08.896832] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.024 [2024-10-17 17:50:08.896845] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.024 [2024-10-17 17:50:08.896852] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.024 [2024-10-17 17:50:08.896859] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.024 [2024-10-17 17:50:08.896873] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.024 qpair failed and we were unable to recover it. 00:29:01.024 [2024-10-17 17:50:08.906885] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.025 [2024-10-17 17:50:08.906940] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.025 [2024-10-17 17:50:08.906953] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.025 [2024-10-17 17:50:08.906964] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.025 [2024-10-17 17:50:08.906970] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.025 [2024-10-17 17:50:08.906984] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.025 qpair failed and we were unable to recover it. 00:29:01.025 [2024-10-17 17:50:08.916906] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.025 [2024-10-17 17:50:08.916958] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.025 [2024-10-17 17:50:08.916971] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.025 [2024-10-17 17:50:08.916978] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.025 [2024-10-17 17:50:08.916984] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.025 [2024-10-17 17:50:08.916999] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.025 qpair failed and we were unable to recover it. 00:29:01.025 [2024-10-17 17:50:08.926903] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.025 [2024-10-17 17:50:08.926958] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.025 [2024-10-17 17:50:08.926971] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.025 [2024-10-17 17:50:08.926978] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.025 [2024-10-17 17:50:08.926985] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.025 [2024-10-17 17:50:08.926998] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.025 qpair failed and we were unable to recover it. 00:29:01.025 [2024-10-17 17:50:08.936894] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.025 [2024-10-17 17:50:08.936942] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.025 [2024-10-17 17:50:08.936955] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.025 [2024-10-17 17:50:08.936962] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.025 [2024-10-17 17:50:08.936968] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.025 [2024-10-17 17:50:08.936982] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.025 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:08.946872] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:08.946925] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:08.946938] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:08.946945] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:08.946951] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:08.946971] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:08.956983] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:08.957074] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:08.957087] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:08.957094] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:08.957100] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:08.957114] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:08.967019] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:08.967070] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:08.967082] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:08.967089] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:08.967095] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:08.967109] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:08.977023] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:08.977069] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:08.977081] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:08.977088] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:08.977095] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:08.977108] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:08.987067] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:08.987115] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:08.987127] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:08.987134] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:08.987140] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:08.987154] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:08.997093] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:08.997142] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:08.997158] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:08.997165] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:08.997172] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:08.997185] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:09.007110] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:09.007166] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:09.007179] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:09.007186] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:09.007192] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:09.007206] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:09.016991] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:09.017055] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:09.017068] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:09.017075] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:09.017081] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:09.017095] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.286 qpair failed and we were unable to recover it. 00:29:01.286 [2024-10-17 17:50:09.027196] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.286 [2024-10-17 17:50:09.027249] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.286 [2024-10-17 17:50:09.027262] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.286 [2024-10-17 17:50:09.027269] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.286 [2024-10-17 17:50:09.027275] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.286 [2024-10-17 17:50:09.027289] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.037202] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.037293] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.037306] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.037313] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.037319] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.037333] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.047224] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.047275] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.047288] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.047295] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.047301] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.047315] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.057237] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.057283] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.057296] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.057303] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.057309] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.057323] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.067194] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.067245] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.067258] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.067265] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.067271] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.067285] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.077306] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.077357] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.077370] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.077377] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.077383] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.077397] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.087349] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.087394] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.087410] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.087417] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.087424] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.087438] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.097344] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.097399] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.097412] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.097419] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.097425] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.097438] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.107416] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.107466] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.107479] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.107486] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.107493] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.107507] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.117387] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.117432] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.117445] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.117452] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.117458] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.117472] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.127454] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.127550] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.127563] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.127570] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.127576] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.127594] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.137445] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.137489] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.137502] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.137509] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.137515] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.137529] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.147510] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.147561] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.147574] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.147581] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.147588] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.147602] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.157526] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.157575] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.157587] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.157595] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.157601] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.157614] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.167609] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.287 [2024-10-17 17:50:09.167673] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.287 [2024-10-17 17:50:09.167685] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.287 [2024-10-17 17:50:09.167698] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.287 [2024-10-17 17:50:09.167705] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.287 [2024-10-17 17:50:09.167720] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.287 qpair failed and we were unable to recover it. 00:29:01.287 [2024-10-17 17:50:09.177428] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.288 [2024-10-17 17:50:09.177475] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.288 [2024-10-17 17:50:09.177491] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.288 [2024-10-17 17:50:09.177499] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.288 [2024-10-17 17:50:09.177505] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.288 [2024-10-17 17:50:09.177519] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.288 qpair failed and we were unable to recover it. 00:29:01.288 [2024-10-17 17:50:09.187630] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.288 [2024-10-17 17:50:09.187679] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.288 [2024-10-17 17:50:09.187697] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.288 [2024-10-17 17:50:09.187704] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.288 [2024-10-17 17:50:09.187710] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.288 [2024-10-17 17:50:09.187725] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.288 qpair failed and we were unable to recover it. 00:29:01.288 [2024-10-17 17:50:09.197659] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.288 [2024-10-17 17:50:09.197711] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.288 [2024-10-17 17:50:09.197724] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.288 [2024-10-17 17:50:09.197731] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.288 [2024-10-17 17:50:09.197737] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.288 [2024-10-17 17:50:09.197751] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.288 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.207649] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.207701] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.207714] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.207721] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.207727] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.207742] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.217651] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.217703] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.217716] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.217723] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.217732] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.217747] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.227723] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.227780] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.227792] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.227799] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.227806] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.227820] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.237756] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.237805] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.237818] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.237825] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.237831] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.237845] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.247789] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.247836] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.247848] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.247855] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.247861] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.247875] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.257780] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.257830] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.257843] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.257850] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.257856] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.257869] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.267839] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.267892] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.267905] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.267912] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.267918] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.267933] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.277865] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.277948] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.277961] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.277968] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.277974] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.277988] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.287881] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.287927] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.287940] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.287947] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.287953] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.287967] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.297744] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.297789] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.297802] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.297809] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.297816] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.297829] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.307934] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.308000] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.308015] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.308022] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.308034] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.308049] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.317974] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.550 [2024-10-17 17:50:09.318025] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.550 [2024-10-17 17:50:09.318039] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.550 [2024-10-17 17:50:09.318046] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.550 [2024-10-17 17:50:09.318052] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.550 [2024-10-17 17:50:09.318066] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.550 qpair failed and we were unable to recover it. 00:29:01.550 [2024-10-17 17:50:09.328094] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.328171] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.328184] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.328192] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.328198] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.328212] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.337983] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.338031] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.338044] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.338051] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.338058] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.338072] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.348049] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.348115] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.348128] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.348135] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.348141] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.348155] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.357934] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.357983] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.357996] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.358003] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.358010] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.358023] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.368066] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.368113] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.368126] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.368133] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.368139] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.368153] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.378077] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.378127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.378139] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.378147] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.378153] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.378166] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.388156] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.388208] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.388221] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.388228] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.388234] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.388248] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.398043] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.398099] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.398112] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.398126] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.398132] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.398146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.408203] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.408248] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.408261] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.408268] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.408275] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.408288] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.418204] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.418251] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.418265] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.418271] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.418278] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.418291] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.428269] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.428321] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.428334] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.428341] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.428347] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.428361] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.438281] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.438332] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.438345] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.438352] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.438358] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.438371] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.448315] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.448369] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.448386] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.448393] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.448399] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.448414] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.551 [2024-10-17 17:50:09.458317] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.551 [2024-10-17 17:50:09.458362] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.551 [2024-10-17 17:50:09.458375] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.551 [2024-10-17 17:50:09.458382] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.551 [2024-10-17 17:50:09.458388] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.551 [2024-10-17 17:50:09.458402] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.551 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.468383] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.468439] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.468463] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.468472] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.468479] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.468498] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.478394] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.478479] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.478503] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.478512] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.478519] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.478538] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.488416] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.488468] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.488492] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.488506] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.488513] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.488532] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.498434] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.498484] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.498499] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.498506] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.498512] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.498527] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.508488] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.508545] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.508559] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.508566] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.508572] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.508586] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.518504] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.518555] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.518569] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.518575] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.518582] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.518596] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.528532] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.528582] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.528595] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.528602] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.528609] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.528623] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.538492] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.538540] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.538554] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.538561] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.538567] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.538581] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.814 [2024-10-17 17:50:09.548601] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.814 [2024-10-17 17:50:09.548658] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.814 [2024-10-17 17:50:09.548671] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.814 [2024-10-17 17:50:09.548678] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.814 [2024-10-17 17:50:09.548685] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.814 [2024-10-17 17:50:09.548702] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.814 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.558498] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.558596] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.558610] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.558617] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.558623] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.558637] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.568673] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.568767] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.568780] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.568787] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.568794] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.568808] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.578658] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.578710] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.578727] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.578734] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.578740] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.578755] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.588706] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.588792] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.588806] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.588812] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.588819] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.588834] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.598727] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.598774] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.598788] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.598795] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.598801] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.598815] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.608745] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.608827] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.608840] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.608847] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.608853] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.608868] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.618793] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.618856] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.618869] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.618876] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.618882] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.618901] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.628849] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.628940] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.628953] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.628960] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.628966] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.628981] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.638835] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.638883] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.638896] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.638904] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.638910] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.638924] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.648860] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.648909] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.648922] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.648929] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.648935] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.648949] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.658841] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.658890] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.658903] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.658910] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.658916] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.658930] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.668941] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.668987] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.669003] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.669010] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.669016] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.669030] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.678952] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.679002] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.679015] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.679021] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.679028] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.679041] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.689001] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.689066] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.815 [2024-10-17 17:50:09.689079] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.815 [2024-10-17 17:50:09.689086] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.815 [2024-10-17 17:50:09.689092] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.815 [2024-10-17 17:50:09.689107] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.815 qpair failed and we were unable to recover it. 00:29:01.815 [2024-10-17 17:50:09.698937] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.815 [2024-10-17 17:50:09.698984] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.816 [2024-10-17 17:50:09.698997] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.816 [2024-10-17 17:50:09.699004] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.816 [2024-10-17 17:50:09.699010] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.816 [2024-10-17 17:50:09.699024] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.816 qpair failed and we were unable to recover it. 00:29:01.816 [2024-10-17 17:50:09.709030] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.816 [2024-10-17 17:50:09.709077] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.816 [2024-10-17 17:50:09.709090] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.816 [2024-10-17 17:50:09.709097] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.816 [2024-10-17 17:50:09.709103] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.816 [2024-10-17 17:50:09.709120] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.816 qpair failed and we were unable to recover it. 00:29:01.816 [2024-10-17 17:50:09.719061] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.816 [2024-10-17 17:50:09.719126] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.816 [2024-10-17 17:50:09.719139] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.816 [2024-10-17 17:50:09.719146] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.816 [2024-10-17 17:50:09.719152] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.816 [2024-10-17 17:50:09.719166] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.816 qpair failed and we were unable to recover it. 00:29:01.816 [2024-10-17 17:50:09.728969] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:01.816 [2024-10-17 17:50:09.729022] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:01.816 [2024-10-17 17:50:09.729035] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:01.816 [2024-10-17 17:50:09.729042] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:01.816 [2024-10-17 17:50:09.729048] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:01.816 [2024-10-17 17:50:09.729062] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:01.816 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.739098] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.739146] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.739160] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.739167] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.739173] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.739187] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.749040] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.749136] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.749149] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.749156] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.749162] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.749176] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.759170] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.759249] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.759262] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.759269] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.759275] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.759289] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.769191] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.769238] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.769251] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.769258] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.769264] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.769278] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.779058] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.779105] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.779118] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.779125] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.779132] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.779145] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.789292] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.789376] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.789389] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.789396] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.789403] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.789416] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.799161] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.799242] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.799265] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.799272] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.799283] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.799303] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.809296] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.809341] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.809354] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.809361] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.078 [2024-10-17 17:50:09.809368] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.078 [2024-10-17 17:50:09.809381] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.078 qpair failed and we were unable to recover it. 00:29:02.078 [2024-10-17 17:50:09.819301] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.078 [2024-10-17 17:50:09.819351] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.078 [2024-10-17 17:50:09.819365] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.078 [2024-10-17 17:50:09.819372] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.819378] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.819391] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.829379] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.829427] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.829441] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.829448] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.829454] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.829468] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.839241] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.839293] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.839306] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.839313] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.839319] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.839333] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.849402] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.849453] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.849466] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.849473] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.849479] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.849493] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.859383] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.859438] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.859462] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.859471] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.859478] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.859497] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.869394] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.869450] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.869474] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.869483] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.869490] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.869509] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.879471] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.879524] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.879549] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.879557] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.879565] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.879583] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.889396] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.889462] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.889477] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.889489] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.889496] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.889511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.899523] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.899596] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.899609] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.899616] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.899623] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.899637] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.909550] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.909603] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.909616] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.909623] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.909629] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.909643] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.919595] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.919645] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.919659] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.919666] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.919672] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.919686] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.929637] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.929688] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.929705] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.929712] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.929718] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.929733] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.939609] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.939665] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.939679] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.939686] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.939696] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.939710] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.949661] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.949710] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.949723] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.949730] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.949737] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.949751] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.079 [2024-10-17 17:50:09.959710] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.079 [2024-10-17 17:50:09.959759] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.079 [2024-10-17 17:50:09.959772] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.079 [2024-10-17 17:50:09.959779] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.079 [2024-10-17 17:50:09.959785] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.079 [2024-10-17 17:50:09.959799] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.079 qpair failed and we were unable to recover it. 00:29:02.080 [2024-10-17 17:50:09.969598] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.080 [2024-10-17 17:50:09.969647] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.080 [2024-10-17 17:50:09.969659] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.080 [2024-10-17 17:50:09.969667] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.080 [2024-10-17 17:50:09.969673] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.080 [2024-10-17 17:50:09.969686] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.080 qpair failed and we were unable to recover it. 00:29:02.080 [2024-10-17 17:50:09.979736] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.080 [2024-10-17 17:50:09.979782] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.080 [2024-10-17 17:50:09.979797] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.080 [2024-10-17 17:50:09.979808] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.080 [2024-10-17 17:50:09.979815] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.080 [2024-10-17 17:50:09.979829] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.080 qpair failed and we were unable to recover it. 00:29:02.080 [2024-10-17 17:50:09.989624] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.080 [2024-10-17 17:50:09.989675] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.080 [2024-10-17 17:50:09.989688] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.080 [2024-10-17 17:50:09.989699] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.080 [2024-10-17 17:50:09.989705] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.080 [2024-10-17 17:50:09.989720] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.080 qpair failed and we were unable to recover it. 00:29:02.342 [2024-10-17 17:50:09.999802] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.342 [2024-10-17 17:50:09.999852] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.342 [2024-10-17 17:50:09.999866] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.342 [2024-10-17 17:50:09.999873] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.342 [2024-10-17 17:50:09.999879] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.342 [2024-10-17 17:50:09.999894] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.342 qpair failed and we were unable to recover it. 00:29:02.342 [2024-10-17 17:50:10.009784] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.342 [2024-10-17 17:50:10.009884] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.342 [2024-10-17 17:50:10.009899] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.342 [2024-10-17 17:50:10.009906] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.342 [2024-10-17 17:50:10.009913] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.342 [2024-10-17 17:50:10.009928] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.342 qpair failed and we were unable to recover it. 00:29:02.342 [2024-10-17 17:50:10.019837] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.342 [2024-10-17 17:50:10.019893] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.342 [2024-10-17 17:50:10.019908] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.342 [2024-10-17 17:50:10.019916] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.342 [2024-10-17 17:50:10.019923] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.342 [2024-10-17 17:50:10.019938] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.342 qpair failed and we were unable to recover it. 00:29:02.342 [2024-10-17 17:50:10.029860] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.342 [2024-10-17 17:50:10.029909] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.342 [2024-10-17 17:50:10.029923] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.342 [2024-10-17 17:50:10.029930] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.342 [2024-10-17 17:50:10.029937] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.029951] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.039816] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.039864] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.039883] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.039891] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.039898] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.039915] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.049988] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.050047] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.050062] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.050070] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.050077] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.050091] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.059954] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.060001] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.060016] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.060023] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.060030] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.060045] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.070016] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.070068] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.070086] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.070094] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.070101] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.070116] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.080049] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.080097] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.080117] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.080126] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.080133] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.080150] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.090065] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.090119] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.090133] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.090140] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.090146] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.090161] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.099958] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.100015] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.100029] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.100036] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.100043] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.100057] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.110143] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.110202] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.110218] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.110225] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.110233] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.110255] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.120012] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.120065] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.120080] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.120088] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.120094] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.120109] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.130155] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.130203] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.130222] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.130234] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.130244] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.130267] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.140067] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.140127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.140146] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.140155] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.140161] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.140178] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.150206] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.150257] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.150271] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.150279] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.150285] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.150300] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.160123] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.160173] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.160195] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.160202] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.160208] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.160223] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.170250] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.343 [2024-10-17 17:50:10.170298] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.343 [2024-10-17 17:50:10.170311] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.343 [2024-10-17 17:50:10.170318] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.343 [2024-10-17 17:50:10.170324] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.343 [2024-10-17 17:50:10.170338] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.343 qpair failed and we were unable to recover it. 00:29:02.343 [2024-10-17 17:50:10.180138] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.180191] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.180205] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.180211] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.180218] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.180232] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.344 [2024-10-17 17:50:10.190295] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.190355] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.190368] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.190375] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.190382] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.190396] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.344 [2024-10-17 17:50:10.200355] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.200401] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.200415] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.200422] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.200428] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.200446] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.344 [2024-10-17 17:50:10.210241] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.210320] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.210333] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.210340] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.210347] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.210361] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.344 [2024-10-17 17:50:10.220359] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.220406] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.220419] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.220426] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.220432] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.220447] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.344 [2024-10-17 17:50:10.230398] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.230443] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.230456] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.230463] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.230470] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.230484] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.344 [2024-10-17 17:50:10.240501] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.240554] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.240567] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.240574] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.240581] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.240595] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.344 [2024-10-17 17:50:10.250454] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.344 [2024-10-17 17:50:10.250503] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.344 [2024-10-17 17:50:10.250519] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.344 [2024-10-17 17:50:10.250527] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.344 [2024-10-17 17:50:10.250533] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.344 [2024-10-17 17:50:10.250547] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.344 qpair failed and we were unable to recover it. 00:29:02.606 [2024-10-17 17:50:10.260468] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.606 [2024-10-17 17:50:10.260516] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.606 [2024-10-17 17:50:10.260529] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.606 [2024-10-17 17:50:10.260537] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.606 [2024-10-17 17:50:10.260543] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.606 [2024-10-17 17:50:10.260558] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.606 qpair failed and we were unable to recover it. 00:29:02.606 [2024-10-17 17:50:10.270370] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.606 [2024-10-17 17:50:10.270422] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.270435] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.270442] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.270448] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.270463] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.280564] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.280611] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.280624] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.280631] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.280637] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.280651] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.290602] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.290652] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.290665] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.290672] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.290682] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.290699] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.300585] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.300634] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.300648] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.300656] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.300663] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.300676] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.310626] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.310672] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.310685] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.310704] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.310711] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.310725] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.320622] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.320670] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.320683] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.320693] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.320700] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.320714] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.330694] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.330739] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.330752] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.330759] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.330765] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.330779] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.340685] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.340740] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.340753] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.340760] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.340766] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.340780] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.350730] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.350801] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.350814] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.350820] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.350827] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.350840] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.360747] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.360795] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.360808] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.360815] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.360821] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.360835] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.370776] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.370817] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.607 [2024-10-17 17:50:10.370829] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.607 [2024-10-17 17:50:10.370836] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.607 [2024-10-17 17:50:10.370842] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.607 [2024-10-17 17:50:10.370856] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.607 qpair failed and we were unable to recover it. 00:29:02.607 [2024-10-17 17:50:10.380806] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.607 [2024-10-17 17:50:10.380853] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.380866] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.380873] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.380883] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.380897] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.390830] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.390873] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.390886] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.390893] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.390899] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.390913] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.400767] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.400824] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.400837] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.400844] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.400850] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.400864] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.410764] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.410811] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.410823] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.410830] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.410837] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.410850] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.420830] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.420875] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.420888] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.420895] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.420901] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.420914] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.430940] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.430990] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.431003] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.431010] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.431016] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.431030] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.440866] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.440912] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.440925] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.440932] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.440938] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.440952] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.450964] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.451041] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.451053] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.451060] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.451066] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.451080] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.461023] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.461068] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.461081] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.461087] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.461094] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.461107] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.471032] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.471083] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.471096] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.471106] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.471112] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.471126] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.481100] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.481150] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.481162] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.481169] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.481175] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.481189] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.491066] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.491114] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.491127] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.491133] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.491140] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.491153] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.501119] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.501181] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.501194] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.501200] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.501207] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.501220] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.511189] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.608 [2024-10-17 17:50:10.511273] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.608 [2024-10-17 17:50:10.511286] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.608 [2024-10-17 17:50:10.511293] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.608 [2024-10-17 17:50:10.511299] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.608 [2024-10-17 17:50:10.511313] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.608 qpair failed and we were unable to recover it. 00:29:02.608 [2024-10-17 17:50:10.521205] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.609 [2024-10-17 17:50:10.521254] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.609 [2024-10-17 17:50:10.521267] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.609 [2024-10-17 17:50:10.521274] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.609 [2024-10-17 17:50:10.521280] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.609 [2024-10-17 17:50:10.521294] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.609 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.531203] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.531256] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.531270] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.531276] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.531283] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.531296] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.541240] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.541286] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.541299] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.541306] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.541312] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.541326] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.551130] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.551179] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.551192] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.551200] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.551207] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.551221] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.561195] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.561240] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.561256] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.561263] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.561270] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.561283] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.571312] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.571359] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.571372] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.571379] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.571386] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.571399] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.581343] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.581391] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.581404] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.581411] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.581417] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.581431] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.591374] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.591417] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.591430] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.591436] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.591443] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.591456] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.601436] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.601491] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.601504] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.601510] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.601517] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.601530] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.611420] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.611463] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.611476] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.611483] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.611490] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.611504] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.621448] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.621495] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.621509] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.621515] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.621522] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.871 [2024-10-17 17:50:10.621535] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.871 qpair failed and we were unable to recover it. 00:29:02.871 [2024-10-17 17:50:10.631462] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.871 [2024-10-17 17:50:10.631511] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.871 [2024-10-17 17:50:10.631525] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.871 [2024-10-17 17:50:10.631532] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.871 [2024-10-17 17:50:10.631538] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.631552] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.641515] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.641627] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.641641] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.641648] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.641654] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.641667] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.651405] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.651445] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.651461] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.651468] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.651474] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.651488] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.661552] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.661643] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.661657] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.661663] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.661670] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.661684] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.671590] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.671636] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.671649] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.671656] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.671662] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.671676] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.681576] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.681631] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.681645] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.681651] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.681657] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.681671] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.691645] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.691693] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.691706] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.691713] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.691719] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.691737] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.701681] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.701736] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.701749] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.701756] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.701763] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.701777] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.711576] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.711622] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.711635] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.711642] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.711648] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.711662] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.721766] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.721842] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.721855] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.721862] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.721868] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.721882] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.731750] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.731798] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.731811] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.731818] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.731824] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.731838] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.741667] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.741718] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.741734] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.741741] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.741747] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.741761] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.751719] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.751771] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.751784] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.751791] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.751797] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.751811] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.761852] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.761898] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.761910] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.761917] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.761924] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.872 [2024-10-17 17:50:10.761937] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.872 qpair failed and we were unable to recover it. 00:29:02.872 [2024-10-17 17:50:10.771829] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.872 [2024-10-17 17:50:10.771872] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.872 [2024-10-17 17:50:10.771885] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.872 [2024-10-17 17:50:10.771892] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.872 [2024-10-17 17:50:10.771898] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.873 [2024-10-17 17:50:10.771912] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.873 qpair failed and we were unable to recover it. 00:29:02.873 [2024-10-17 17:50:10.781908] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:02.873 [2024-10-17 17:50:10.781952] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:02.873 [2024-10-17 17:50:10.781964] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:02.873 [2024-10-17 17:50:10.781971] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:02.873 [2024-10-17 17:50:10.781981] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:02.873 [2024-10-17 17:50:10.781994] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:02.873 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.791941] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.791990] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.792003] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.792010] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.792016] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.792030] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.801934] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.801981] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.801994] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.802001] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.802007] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.802021] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.811840] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.811885] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.811898] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.811905] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.811911] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.811925] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.821981] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.822025] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.822038] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.822045] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.822051] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.822065] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.832043] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.832093] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.832106] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.832113] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.832119] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.832133] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.842050] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.842097] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.842110] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.842117] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.842123] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.842137] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.852068] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.852111] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.852124] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.852130] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.852136] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.852150] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.861968] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.862013] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.862026] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.862032] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.862039] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.862052] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.872015] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.872064] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.872077] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.872084] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.872093] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.872107] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.135 qpair failed and we were unable to recover it. 00:29:03.135 [2024-10-17 17:50:10.882022] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.135 [2024-10-17 17:50:10.882064] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.135 [2024-10-17 17:50:10.882078] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.135 [2024-10-17 17:50:10.882084] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.135 [2024-10-17 17:50:10.882091] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.135 [2024-10-17 17:50:10.882105] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.892192] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.892239] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.892252] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.892258] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.892265] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.892279] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.902224] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.902272] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.902285] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.902292] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.902298] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.902313] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.912139] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.912203] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.912216] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.912223] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.912229] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.912242] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.922267] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.922331] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.922344] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.922351] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.922357] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.922371] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.932265] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.932308] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.932321] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.932328] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.932334] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.932348] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.942325] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.942410] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.942423] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.942429] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.942435] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.942449] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.952355] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.952418] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.952443] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.952451] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.952458] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.952477] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.962368] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.962413] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.962428] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.962439] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.962446] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.962461] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.972401] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.972453] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.972477] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.972486] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.972492] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.972511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.982443] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.982494] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.982519] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.982527] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.982534] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.982552] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:10.992494] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:10.992543] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:10.992557] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:10.992565] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:10.992571] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:10.992586] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.136 qpair failed and we were unable to recover it. 00:29:03.136 [2024-10-17 17:50:11.002495] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.136 [2024-10-17 17:50:11.002539] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.136 [2024-10-17 17:50:11.002552] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.136 [2024-10-17 17:50:11.002559] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.136 [2024-10-17 17:50:11.002566] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.136 [2024-10-17 17:50:11.002580] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.137 qpair failed and we were unable to recover it. 00:29:03.137 [2024-10-17 17:50:11.012375] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.137 [2024-10-17 17:50:11.012422] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.137 [2024-10-17 17:50:11.012435] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.137 [2024-10-17 17:50:11.012442] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.137 [2024-10-17 17:50:11.012449] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.137 [2024-10-17 17:50:11.012463] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.137 qpair failed and we were unable to recover it. 00:29:03.137 [2024-10-17 17:50:11.022592] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.137 [2024-10-17 17:50:11.022643] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.137 [2024-10-17 17:50:11.022656] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.137 [2024-10-17 17:50:11.022663] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.137 [2024-10-17 17:50:11.022669] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.137 [2024-10-17 17:50:11.022683] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.137 qpair failed and we were unable to recover it. 00:29:03.137 [2024-10-17 17:50:11.032596] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.137 [2024-10-17 17:50:11.032648] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.137 [2024-10-17 17:50:11.032661] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.137 [2024-10-17 17:50:11.032668] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.137 [2024-10-17 17:50:11.032675] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.137 [2024-10-17 17:50:11.032688] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.137 qpair failed and we were unable to recover it. 00:29:03.137 [2024-10-17 17:50:11.042602] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.137 [2024-10-17 17:50:11.042645] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.137 [2024-10-17 17:50:11.042658] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.137 [2024-10-17 17:50:11.042665] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.137 [2024-10-17 17:50:11.042671] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.137 [2024-10-17 17:50:11.042685] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.137 qpair failed and we were unable to recover it. 00:29:03.137 [2024-10-17 17:50:11.052611] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.399 [2024-10-17 17:50:11.052652] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.399 [2024-10-17 17:50:11.052666] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.399 [2024-10-17 17:50:11.052678] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.399 [2024-10-17 17:50:11.052686] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.399 [2024-10-17 17:50:11.052706] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.399 qpair failed and we were unable to recover it. 00:29:03.399 [2024-10-17 17:50:11.062518] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.399 [2024-10-17 17:50:11.062571] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.399 [2024-10-17 17:50:11.062584] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.399 [2024-10-17 17:50:11.062591] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.399 [2024-10-17 17:50:11.062598] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.399 [2024-10-17 17:50:11.062612] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.399 qpair failed and we were unable to recover it. 00:29:03.399 [2024-10-17 17:50:11.072696] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.399 [2024-10-17 17:50:11.072753] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.399 [2024-10-17 17:50:11.072767] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.399 [2024-10-17 17:50:11.072774] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.399 [2024-10-17 17:50:11.072780] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.399 [2024-10-17 17:50:11.072795] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.399 qpair failed and we were unable to recover it. 00:29:03.399 [2024-10-17 17:50:11.082699] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.399 [2024-10-17 17:50:11.082743] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.399 [2024-10-17 17:50:11.082757] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.399 [2024-10-17 17:50:11.082766] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.399 [2024-10-17 17:50:11.082773] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.399 [2024-10-17 17:50:11.082788] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.399 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.092739] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.092795] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.092808] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.092815] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.092822] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.092836] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.102749] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.102795] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.102808] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.102815] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.102822] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.102836] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.112799] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.112847] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.112860] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.112867] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.112874] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.112888] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.122719] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.122765] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.122778] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.122785] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.122792] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.122806] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.132713] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.132762] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.132775] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.132782] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.132789] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.132803] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.142886] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.142931] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.142949] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.142956] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.142963] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.142977] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.152771] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.152818] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.152832] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.152838] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.152845] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.152859] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.162809] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.162860] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.162874] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.162881] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.162888] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.162902] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.172968] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.173012] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.173025] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.173032] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.173038] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.173052] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.182979] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.183033] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.183046] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.183053] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.183060] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.183077] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.193015] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.193065] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.193078] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.193086] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.193092] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.193107] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.203029] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.203071] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.203084] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.203092] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.203098] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.203112] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.213054] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.213139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.213152] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.213160] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.213167] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.213182] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.223093] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.400 [2024-10-17 17:50:11.223141] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.400 [2024-10-17 17:50:11.223154] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.400 [2024-10-17 17:50:11.223161] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.400 [2024-10-17 17:50:11.223168] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.400 [2024-10-17 17:50:11.223182] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.400 qpair failed and we were unable to recover it. 00:29:03.400 [2024-10-17 17:50:11.233128] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.233180] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.233197] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.233204] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.233210] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.233225] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.243132] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.243175] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.243189] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.243196] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.243203] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.243217] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.253158] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.253201] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.253214] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.253222] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.253229] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.253243] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.263204] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.263288] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.263301] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.263308] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.263315] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.263329] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.273212] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.273260] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.273273] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.273280] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.273290] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.273304] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.283148] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.283197] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.283210] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.283217] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.283224] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.283237] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.293133] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.293178] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.293191] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.293198] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.293204] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.293218] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.303319] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.303375] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.303388] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.303395] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.303402] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.303417] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.401 [2024-10-17 17:50:11.313343] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.401 [2024-10-17 17:50:11.313422] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.401 [2024-10-17 17:50:11.313436] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.401 [2024-10-17 17:50:11.313444] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.401 [2024-10-17 17:50:11.313450] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.401 [2024-10-17 17:50:11.313464] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.401 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.323347] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.323393] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.323406] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.323413] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.323419] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.323433] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.333421] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.333470] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.333483] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.333490] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.333497] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.333511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.343379] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.343427] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.343440] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.343448] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.343454] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.343468] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.353320] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.353366] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.353380] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.353387] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.353393] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.353409] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.363336] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.363384] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.363399] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.363410] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.363423] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.363440] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.373489] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.373534] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.373547] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.373557] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.373564] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.373578] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.383520] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.383568] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.383582] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.383589] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.383595] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.383609] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.393559] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.393603] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.393616] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.393623] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.393629] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.393643] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.403438] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.403481] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.403494] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.403501] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.403507] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.664 [2024-10-17 17:50:11.403521] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.664 qpair failed and we were unable to recover it. 00:29:03.664 [2024-10-17 17:50:11.413586] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.664 [2024-10-17 17:50:11.413641] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.664 [2024-10-17 17:50:11.413655] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.664 [2024-10-17 17:50:11.413662] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.664 [2024-10-17 17:50:11.413668] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.413683] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.423490] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.423540] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.423553] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.423560] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.423566] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.423579] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.433650] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.433709] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.433723] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.433730] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.433745] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.433759] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.443673] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.443720] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.443734] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.443741] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.443748] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.443763] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.453573] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.453623] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.453636] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.453646] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.453653] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.453667] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.463740] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.463796] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.463809] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.463816] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.463823] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.463836] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.473634] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.473679] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.473695] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.473702] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.473708] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.473722] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.483743] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.483788] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.483801] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.483808] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.483814] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.483828] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.493684] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.493736] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.493750] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.493757] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.493763] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.493777] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.503716] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.503764] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.503777] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.503784] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.503790] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.503804] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.513889] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.513938] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.513952] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.513959] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.513965] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.513979] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.523885] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.523927] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.523940] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.523947] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.523954] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.523968] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.533901] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.533946] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.533959] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.533966] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.533973] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.533986] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.543829] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.543876] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.543888] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.543906] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.543912] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.543926] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.665 qpair failed and we were unable to recover it. 00:29:03.665 [2024-10-17 17:50:11.553980] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.665 [2024-10-17 17:50:11.554042] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.665 [2024-10-17 17:50:11.554055] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.665 [2024-10-17 17:50:11.554062] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.665 [2024-10-17 17:50:11.554068] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.665 [2024-10-17 17:50:11.554082] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.666 qpair failed and we were unable to recover it. 00:29:03.666 [2024-10-17 17:50:11.563881] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.666 [2024-10-17 17:50:11.563929] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.666 [2024-10-17 17:50:11.563942] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.666 [2024-10-17 17:50:11.563949] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.666 [2024-10-17 17:50:11.563955] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.666 [2024-10-17 17:50:11.563970] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.666 qpair failed and we were unable to recover it. 00:29:03.666 [2024-10-17 17:50:11.573908] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.666 [2024-10-17 17:50:11.573974] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.666 [2024-10-17 17:50:11.573987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.666 [2024-10-17 17:50:11.573993] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.666 [2024-10-17 17:50:11.573999] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.666 [2024-10-17 17:50:11.574013] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.666 qpair failed and we were unable to recover it. 00:29:03.928 [2024-10-17 17:50:11.584073] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.928 [2024-10-17 17:50:11.584121] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.928 [2024-10-17 17:50:11.584134] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.928 [2024-10-17 17:50:11.584141] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.928 [2024-10-17 17:50:11.584147] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.928 [2024-10-17 17:50:11.584161] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.928 qpair failed and we were unable to recover it. 00:29:03.928 [2024-10-17 17:50:11.594100] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.928 [2024-10-17 17:50:11.594152] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.928 [2024-10-17 17:50:11.594166] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.928 [2024-10-17 17:50:11.594173] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.928 [2024-10-17 17:50:11.594179] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.928 [2024-10-17 17:50:11.594197] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.928 qpair failed and we were unable to recover it. 00:29:03.928 [2024-10-17 17:50:11.604142] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.928 [2024-10-17 17:50:11.604182] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.928 [2024-10-17 17:50:11.604195] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.928 [2024-10-17 17:50:11.604202] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.928 [2024-10-17 17:50:11.604208] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.928 [2024-10-17 17:50:11.604222] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.928 qpair failed and we were unable to recover it. 00:29:03.928 [2024-10-17 17:50:11.614134] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.928 [2024-10-17 17:50:11.614219] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.928 [2024-10-17 17:50:11.614233] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.928 [2024-10-17 17:50:11.614240] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.928 [2024-10-17 17:50:11.614246] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.928 [2024-10-17 17:50:11.614260] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.928 qpair failed and we were unable to recover it. 00:29:03.928 [2024-10-17 17:50:11.624161] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.928 [2024-10-17 17:50:11.624210] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.928 [2024-10-17 17:50:11.624223] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.928 [2024-10-17 17:50:11.624230] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.928 [2024-10-17 17:50:11.624236] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.928 [2024-10-17 17:50:11.624250] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.928 qpair failed and we were unable to recover it. 00:29:03.928 [2024-10-17 17:50:11.634134] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.928 [2024-10-17 17:50:11.634183] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.634200] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.634207] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.634213] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.634227] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.644222] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.644264] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.644277] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.644284] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.644290] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.644304] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.654240] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.654292] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.654305] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.654312] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.654318] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.654332] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.664191] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.664255] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.664268] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.664274] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.664281] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.664295] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.674290] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.674340] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.674353] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.674360] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.674366] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.674387] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.684336] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.684385] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.684398] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.684405] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.684411] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.684425] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.694361] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.694453] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.694466] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.694474] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.694480] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.694493] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.704388] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.704434] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.704448] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.704455] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.704461] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.704475] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.714438] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.714488] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.714502] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.714509] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.714516] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.714530] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.724401] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.724473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.724489] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.724496] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.724502] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.724516] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.734474] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.734516] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.734529] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.734536] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.734542] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.734555] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.744460] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.744507] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.744519] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.744526] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.744533] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.744546] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.754532] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.754581] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.754594] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.754601] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.754607] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.754621] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.764537] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.764583] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.764596] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.764603] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.929 [2024-10-17 17:50:11.764609] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.929 [2024-10-17 17:50:11.764628] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.929 qpair failed and we were unable to recover it. 00:29:03.929 [2024-10-17 17:50:11.774565] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.929 [2024-10-17 17:50:11.774619] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.929 [2024-10-17 17:50:11.774632] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.929 [2024-10-17 17:50:11.774639] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.774645] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.774659] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:03.930 [2024-10-17 17:50:11.784577] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.930 [2024-10-17 17:50:11.784626] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.930 [2024-10-17 17:50:11.784639] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.930 [2024-10-17 17:50:11.784646] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.784652] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.784666] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:03.930 [2024-10-17 17:50:11.794640] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.930 [2024-10-17 17:50:11.794684] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.930 [2024-10-17 17:50:11.794700] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.930 [2024-10-17 17:50:11.794708] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.794714] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.794728] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:03.930 [2024-10-17 17:50:11.804653] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.930 [2024-10-17 17:50:11.804702] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.930 [2024-10-17 17:50:11.804715] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.930 [2024-10-17 17:50:11.804722] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.804729] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.804743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:03.930 [2024-10-17 17:50:11.814669] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.930 [2024-10-17 17:50:11.814722] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.930 [2024-10-17 17:50:11.814739] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.930 [2024-10-17 17:50:11.814747] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.814754] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.814768] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:03.930 [2024-10-17 17:50:11.824702] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.930 [2024-10-17 17:50:11.824754] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.930 [2024-10-17 17:50:11.824767] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.930 [2024-10-17 17:50:11.824773] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.824780] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.824794] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:03.930 [2024-10-17 17:50:11.834747] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.930 [2024-10-17 17:50:11.834794] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.930 [2024-10-17 17:50:11.834807] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.930 [2024-10-17 17:50:11.834814] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.834820] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.834834] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:03.930 [2024-10-17 17:50:11.844728] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:03.930 [2024-10-17 17:50:11.844771] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:03.930 [2024-10-17 17:50:11.844784] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:03.930 [2024-10-17 17:50:11.844790] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:03.930 [2024-10-17 17:50:11.844797] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:03.930 [2024-10-17 17:50:11.844810] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:03.930 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.854767] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.854812] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.854826] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.854833] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.854844] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.854858] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.864833] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.864883] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.864896] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.864903] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.864909] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.864923] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.874859] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.874908] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.874922] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.874929] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.874936] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.874950] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.884884] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.884928] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.884941] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.884949] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.884955] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.884969] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.894873] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.894913] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.894926] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.894933] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.894939] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.894953] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.904914] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.904967] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.904980] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.904987] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.904994] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.905007] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.914842] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.914887] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.914901] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.914908] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.914914] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.914928] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.924972] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.925015] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.925028] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.925035] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.925041] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.925055] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.934981] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.935055] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.935068] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.935075] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.935081] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.935095] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.945037] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.945084] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.945097] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.945108] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.945114] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.945128] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.955072] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.955122] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.955135] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.955142] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.955148] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.955162] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.965089] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.965130] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.965143] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.965150] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.965156] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.965170] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.975106] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.975148] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.975161] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.975167] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.975174] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.193 [2024-10-17 17:50:11.975187] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.193 qpair failed and we were unable to recover it. 00:29:04.193 [2024-10-17 17:50:11.985140] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.193 [2024-10-17 17:50:11.985194] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.193 [2024-10-17 17:50:11.985207] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.193 [2024-10-17 17:50:11.985214] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.193 [2024-10-17 17:50:11.985220] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:11.985234] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:11.995158] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:11.995207] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:11.995220] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:11.995227] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:11.995233] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:11.995247] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.005183] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.005227] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.005240] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.005247] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.005253] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.005267] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.015225] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.015269] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.015282] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.015289] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.015296] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.015309] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.025219] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.025269] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.025282] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.025289] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.025295] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.025308] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.035281] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.035329] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.035343] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.035353] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.035359] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.035373] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.045163] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.045209] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.045222] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.045229] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.045235] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.045249] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.055332] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.055378] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.055391] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.055398] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.055405] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.055419] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.065336] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.065382] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.065396] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.065403] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.065410] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.065423] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.075353] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.075404] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.075417] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.075424] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.075430] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.075443] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.085425] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.085471] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.085484] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.085491] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.085497] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.085511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.095430] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.095473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.095487] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.095493] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.095500] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.095513] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.194 [2024-10-17 17:50:12.105457] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.194 [2024-10-17 17:50:12.105503] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.194 [2024-10-17 17:50:12.105517] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.194 [2024-10-17 17:50:12.105523] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.194 [2024-10-17 17:50:12.105530] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.194 [2024-10-17 17:50:12.105543] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.194 qpair failed and we were unable to recover it. 00:29:04.456 [2024-10-17 17:50:12.115499] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.456 [2024-10-17 17:50:12.115569] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.456 [2024-10-17 17:50:12.115582] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.456 [2024-10-17 17:50:12.115590] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.456 [2024-10-17 17:50:12.115596] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.456 [2024-10-17 17:50:12.115610] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.456 qpair failed and we were unable to recover it. 00:29:04.456 [2024-10-17 17:50:12.125486] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.456 [2024-10-17 17:50:12.125528] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.456 [2024-10-17 17:50:12.125545] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.456 [2024-10-17 17:50:12.125552] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.456 [2024-10-17 17:50:12.125558] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.456 [2024-10-17 17:50:12.125571] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.456 qpair failed and we were unable to recover it. 00:29:04.456 [2024-10-17 17:50:12.135422] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.456 [2024-10-17 17:50:12.135491] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.456 [2024-10-17 17:50:12.135504] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.456 [2024-10-17 17:50:12.135511] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.456 [2024-10-17 17:50:12.135517] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.456 [2024-10-17 17:50:12.135531] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.456 qpair failed and we were unable to recover it. 00:29:04.456 [2024-10-17 17:50:12.145738] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.456 [2024-10-17 17:50:12.145793] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.456 [2024-10-17 17:50:12.145807] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.456 [2024-10-17 17:50:12.145813] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.456 [2024-10-17 17:50:12.145820] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.456 [2024-10-17 17:50:12.145833] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.456 qpair failed and we were unable to recover it. 00:29:04.456 [2024-10-17 17:50:12.155468] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.155546] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.155559] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.155566] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.155572] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.155586] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.165633] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.165674] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.165687] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.165698] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.165705] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.165722] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.175649] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.175697] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.175710] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.175717] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.175723] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.175737] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.185668] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.185718] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.185731] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.185738] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.185744] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.185759] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.195713] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.195763] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.195776] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.195783] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.195789] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.195803] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.205701] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.205742] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.205755] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.205763] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.205769] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.205783] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.215779] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.215872] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.215889] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.215896] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.215902] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.215917] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.225789] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.225836] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.225849] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.225856] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.225862] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.225876] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.235812] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.235858] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.235872] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.235879] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.235885] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.235899] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.245833] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.245904] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.245918] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.245925] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.245931] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.457 [2024-10-17 17:50:12.245949] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.457 qpair failed and we were unable to recover it. 00:29:04.457 [2024-10-17 17:50:12.255745] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.457 [2024-10-17 17:50:12.255793] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.457 [2024-10-17 17:50:12.255806] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.457 [2024-10-17 17:50:12.255813] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.457 [2024-10-17 17:50:12.255820] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.255838] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.265934] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.265980] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.265994] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.266000] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.266007] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.266020] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.275837] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.275887] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.275900] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.275907] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.275913] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.275927] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.285914] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.285993] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.286006] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.286012] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.286019] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.286032] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.296020] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.296065] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.296078] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.296084] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.296091] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.296104] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.305985] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.306032] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.306049] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.306056] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.306062] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.306076] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.316027] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.316079] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.316093] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.316100] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.316106] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.316120] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.326063] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.326108] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.326121] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.326128] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.326134] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.326148] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.336043] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.336087] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.336100] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.336107] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.336114] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.336127] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.346145] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.346213] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.346226] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.346233] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.346243] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.346257] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.356174] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.356225] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.356238] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.356245] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.356251] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.356265] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.458 [2024-10-17 17:50:12.366124] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.458 [2024-10-17 17:50:12.366170] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.458 [2024-10-17 17:50:12.366182] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.458 [2024-10-17 17:50:12.366189] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.458 [2024-10-17 17:50:12.366196] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.458 [2024-10-17 17:50:12.366210] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.458 qpair failed and we were unable to recover it. 00:29:04.721 [2024-10-17 17:50:12.376173] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.721 [2024-10-17 17:50:12.376243] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.721 [2024-10-17 17:50:12.376256] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.721 [2024-10-17 17:50:12.376263] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.721 [2024-10-17 17:50:12.376270] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.721 [2024-10-17 17:50:12.376283] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.721 qpair failed and we were unable to recover it. 00:29:04.721 [2024-10-17 17:50:12.386193] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.721 [2024-10-17 17:50:12.386235] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.721 [2024-10-17 17:50:12.386249] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.721 [2024-10-17 17:50:12.386256] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.721 [2024-10-17 17:50:12.386262] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.721 [2024-10-17 17:50:12.386276] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.721 qpair failed and we were unable to recover it. 00:29:04.721 [2024-10-17 17:50:12.396249] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.721 [2024-10-17 17:50:12.396298] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.721 [2024-10-17 17:50:12.396311] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.721 [2024-10-17 17:50:12.396318] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.721 [2024-10-17 17:50:12.396324] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.396338] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.406277] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.406395] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.406408] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.406416] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.406422] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.406435] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.416277] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.416328] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.416353] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.416361] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.416368] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.416387] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.426275] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.426323] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.426337] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.426344] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.426351] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.426366] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.436357] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.436407] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.436420] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.436428] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.436443] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.436457] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.446354] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.446403] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.446428] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.446437] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.446444] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.446463] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.456402] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.456453] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.456477] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.456486] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.456493] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.456512] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.466413] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.466464] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.466488] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.466497] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.466504] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.466522] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.476454] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.476503] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.476517] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.476524] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.476531] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.476546] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.486466] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.486512] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.486538] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.486546] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.486553] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.486572] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.496501] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.496543] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.496557] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.496565] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.496571] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.496586] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.506540] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.506588] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.506601] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.506608] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.506615] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.506629] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.516595] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.722 [2024-10-17 17:50:12.516687] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.722 [2024-10-17 17:50:12.516704] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.722 [2024-10-17 17:50:12.516710] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.722 [2024-10-17 17:50:12.516717] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.722 [2024-10-17 17:50:12.516731] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.722 qpair failed and we were unable to recover it. 00:29:04.722 [2024-10-17 17:50:12.526570] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.526610] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.526623] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.526635] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.526641] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.526656] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.536572] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.536613] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.536626] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.536633] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.536639] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.536653] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.546516] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.546572] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.546585] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.546592] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.546599] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.546612] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.556681] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.556729] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.556742] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.556749] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.556755] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.556769] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.566688] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.566736] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.566749] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.566755] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.566762] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.566776] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.576712] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.576752] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.576766] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.576772] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.576779] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.576792] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.586735] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.586825] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.586837] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.586844] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.586850] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.586865] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.596778] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.596828] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.596841] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.596847] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.596854] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.596867] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.606777] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.606822] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.606835] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.606842] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.606848] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.606862] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.616796] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.616840] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.616853] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.616864] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.616870] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.616884] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.626719] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.626766] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.626779] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.626786] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.626792] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.626806] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.723 [2024-10-17 17:50:12.636772] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.723 [2024-10-17 17:50:12.636819] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.723 [2024-10-17 17:50:12.636832] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.723 [2024-10-17 17:50:12.636839] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.723 [2024-10-17 17:50:12.636845] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.723 [2024-10-17 17:50:12.636859] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.723 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.646906] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.646948] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.646961] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.646968] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.646974] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.646988] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.656892] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.656936] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.656949] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.656956] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.656962] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.656976] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.666977] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.667053] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.667066] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.667073] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.667079] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.667093] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.676949] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.677001] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.677014] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.677021] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.677027] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.677041] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.686971] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.687012] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.687025] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.687032] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.687038] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.687052] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.697021] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.697061] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.697075] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.697082] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.697088] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.697102] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.707037] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.707081] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.707097] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.707104] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.707110] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.707124] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.717065] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.717129] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.717142] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.717148] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.717155] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.717168] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.727078] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.727124] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.727137] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.727144] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.727150] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.727164] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.737134] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.737208] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.737222] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.737228] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.737234] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.737248] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.747153] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.747228] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.747241] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.747247] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.747254] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.747271] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.757200] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.986 [2024-10-17 17:50:12.757249] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.986 [2024-10-17 17:50:12.757262] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.986 [2024-10-17 17:50:12.757269] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.986 [2024-10-17 17:50:12.757275] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.986 [2024-10-17 17:50:12.757288] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.986 qpair failed and we were unable to recover it. 00:29:04.986 [2024-10-17 17:50:12.767256] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.767307] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.767320] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.767326] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.767333] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.767347] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.777274] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.777363] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.777376] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.777383] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.777389] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.777403] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.787279] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.787323] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.787336] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.787343] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.787349] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.787363] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.797298] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.797348] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.797367] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.797375] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.797381] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.797395] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.807169] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.807213] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.807225] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.807232] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.807238] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.807252] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.817349] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.817390] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.817404] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.817412] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.817418] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.817432] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.827354] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.827406] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.827430] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.827439] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.827446] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.827464] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.837450] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.837540] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.837565] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.837573] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.837585] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.837604] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.847459] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.847505] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.847519] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.847527] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.847533] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.847548] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.857452] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.857536] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.857550] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.857557] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.857563] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.857577] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.867467] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.987 [2024-10-17 17:50:12.867566] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.987 [2024-10-17 17:50:12.867579] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.987 [2024-10-17 17:50:12.867586] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.987 [2024-10-17 17:50:12.867592] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.987 [2024-10-17 17:50:12.867607] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.987 qpair failed and we were unable to recover it. 00:29:04.987 [2024-10-17 17:50:12.877510] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.988 [2024-10-17 17:50:12.877558] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.988 [2024-10-17 17:50:12.877571] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.988 [2024-10-17 17:50:12.877578] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.988 [2024-10-17 17:50:12.877585] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.988 [2024-10-17 17:50:12.877598] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.988 qpair failed and we were unable to recover it. 00:29:04.988 [2024-10-17 17:50:12.887524] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.988 [2024-10-17 17:50:12.887574] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.988 [2024-10-17 17:50:12.887587] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.988 [2024-10-17 17:50:12.887594] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.988 [2024-10-17 17:50:12.887600] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.988 [2024-10-17 17:50:12.887614] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.988 qpair failed and we were unable to recover it. 00:29:04.988 [2024-10-17 17:50:12.897415] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:04.988 [2024-10-17 17:50:12.897457] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:04.988 [2024-10-17 17:50:12.897471] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:04.988 [2024-10-17 17:50:12.897477] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:04.988 [2024-10-17 17:50:12.897484] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:04.988 [2024-10-17 17:50:12.897497] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:04.988 qpair failed and we were unable to recover it. 00:29:05.250 [2024-10-17 17:50:12.907547] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.250 [2024-10-17 17:50:12.907636] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.250 [2024-10-17 17:50:12.907649] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.250 [2024-10-17 17:50:12.907656] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.250 [2024-10-17 17:50:12.907663] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.250 [2024-10-17 17:50:12.907677] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.250 qpair failed and we were unable to recover it. 00:29:05.250 [2024-10-17 17:50:12.917613] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.250 [2024-10-17 17:50:12.917656] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.250 [2024-10-17 17:50:12.917669] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.250 [2024-10-17 17:50:12.917676] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.250 [2024-10-17 17:50:12.917682] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.250 [2024-10-17 17:50:12.917700] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.250 qpair failed and we were unable to recover it. 00:29:05.250 [2024-10-17 17:50:12.927636] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.250 [2024-10-17 17:50:12.927677] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.250 [2024-10-17 17:50:12.927693] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.250 [2024-10-17 17:50:12.927700] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.250 [2024-10-17 17:50:12.927711] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.250 [2024-10-17 17:50:12.927725] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.250 qpair failed and we were unable to recover it. 00:29:05.250 [2024-10-17 17:50:12.937661] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.250 [2024-10-17 17:50:12.937704] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.250 [2024-10-17 17:50:12.937717] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.250 [2024-10-17 17:50:12.937724] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.250 [2024-10-17 17:50:12.937730] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.250 [2024-10-17 17:50:12.937744] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.250 qpair failed and we were unable to recover it. 00:29:05.250 [2024-10-17 17:50:12.947694] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:12.947738] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:12.947751] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:12.947757] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:12.947764] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:12.947778] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:12.957634] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:12.957705] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:12.957719] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:12.957727] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:12.957734] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:12.957753] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:12.967747] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:12.967819] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:12.967833] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:12.967840] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:12.967846] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:12.967860] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:12.977772] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:12.977822] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:12.977835] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:12.977842] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:12.977849] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:12.977862] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:12.987809] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:12.987861] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:12.987874] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:12.987881] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:12.987887] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:12.987901] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:12.997836] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:12.997884] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:12.997897] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:12.997904] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:12.997910] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:12.997924] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.007748] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.007790] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.007803] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.007811] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.007817] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.007831] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.017836] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.017882] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.017895] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.017906] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.017912] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.017926] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.027905] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.027951] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.027964] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.027971] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.027977] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.027991] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.037939] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.037987] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.037999] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.038006] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.038012] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.038026] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.047976] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.048048] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.048061] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.048068] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.048074] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.048088] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.058030] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.058099] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.058112] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.058119] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.058125] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.058139] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.068038] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.068085] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.068098] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.068106] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.068113] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.068128] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.251 [2024-10-17 17:50:13.078045] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.251 [2024-10-17 17:50:13.078116] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.251 [2024-10-17 17:50:13.078129] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.251 [2024-10-17 17:50:13.078136] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.251 [2024-10-17 17:50:13.078142] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.251 [2024-10-17 17:50:13.078156] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.251 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.087917] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.087959] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.087972] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.087979] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.087985] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.087999] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.098097] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.098143] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.098156] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.098163] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.098169] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.098182] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.108113] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.108160] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.108173] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.108183] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.108189] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.108203] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.118154] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.118252] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.118265] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.118272] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.118278] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.118292] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.128153] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.128200] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.128213] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.128220] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.128226] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.128240] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.138232] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.138302] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.138315] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.138322] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.138328] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.138341] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.148222] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.148268] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.148281] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.148288] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.148294] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.148308] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.252 [2024-10-17 17:50:13.158259] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.252 [2024-10-17 17:50:13.158334] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.252 [2024-10-17 17:50:13.158348] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.252 [2024-10-17 17:50:13.158355] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.252 [2024-10-17 17:50:13.158363] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.252 [2024-10-17 17:50:13.158378] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.252 qpair failed and we were unable to recover it. 00:29:05.515 [2024-10-17 17:50:13.168275] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.515 [2024-10-17 17:50:13.168323] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.515 [2024-10-17 17:50:13.168336] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.515 [2024-10-17 17:50:13.168343] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.515 [2024-10-17 17:50:13.168350] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.515 [2024-10-17 17:50:13.168364] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.515 qpair failed and we were unable to recover it. 00:29:05.515 [2024-10-17 17:50:13.178295] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.515 [2024-10-17 17:50:13.178335] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.515 [2024-10-17 17:50:13.178344] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.515 [2024-10-17 17:50:13.178349] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.515 [2024-10-17 17:50:13.178353] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.515 [2024-10-17 17:50:13.178363] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.515 qpair failed and we were unable to recover it. 00:29:05.515 [2024-10-17 17:50:13.188310] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.515 [2024-10-17 17:50:13.188355] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.515 [2024-10-17 17:50:13.188364] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.515 [2024-10-17 17:50:13.188369] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.188373] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.188383] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.198232] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.198277] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.198292] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.198297] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.198302] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.198311] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.208379] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.208417] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.208426] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.208431] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.208435] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.208445] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.218394] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.218442] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.218460] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.218466] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.218471] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.218484] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.228432] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.228477] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.228496] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.228502] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.228507] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.228520] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.238462] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.238507] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.238525] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.238531] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.238536] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.238553] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.248476] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.248516] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.248535] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.248541] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.248546] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.248559] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.258490] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.258530] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.258541] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.258546] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.258550] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.258561] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.268564] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.268654] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.268664] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.268669] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.268673] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.268683] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.278555] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.278598] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.278608] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.278613] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.278617] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.278627] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.288575] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.288616] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.288629] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.288634] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.288638] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.288649] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.298523] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.298582] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.298593] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.298598] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.298602] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.298613] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.308653] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.308696] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.308706] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.308711] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.308716] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.308726] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.318695] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.318744] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.318755] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.318759] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.516 [2024-10-17 17:50:13.318764] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.516 [2024-10-17 17:50:13.318775] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.516 qpair failed and we were unable to recover it. 00:29:05.516 [2024-10-17 17:50:13.328709] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.516 [2024-10-17 17:50:13.328743] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.516 [2024-10-17 17:50:13.328752] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.516 [2024-10-17 17:50:13.328757] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.328762] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.328775] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.338583] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.338623] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.338632] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.338637] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.338642] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.338651] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.348626] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.348667] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.348676] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.348681] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.348685] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.348698] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.358801] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.358847] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.358857] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.358862] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.358866] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.358876] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.368781] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.368817] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.368827] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.368831] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.368836] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.368846] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.378834] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.378913] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.378925] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.378930] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.378934] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.378944] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.388729] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.388790] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.388799] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.388804] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.388808] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.388818] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.398763] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.398838] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.398848] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.398853] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.398857] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.398867] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.408928] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.408994] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.409004] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.409009] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.409013] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.409023] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.418795] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.418834] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.418844] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.418849] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.418856] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.418866] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.517 [2024-10-17 17:50:13.428973] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.517 [2024-10-17 17:50:13.429023] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.517 [2024-10-17 17:50:13.429033] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.517 [2024-10-17 17:50:13.429038] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.517 [2024-10-17 17:50:13.429042] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.517 [2024-10-17 17:50:13.429052] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.517 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.439009] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.439059] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.439069] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.439074] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.439078] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.779 [2024-10-17 17:50:13.439088] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.449025] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.449064] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.449074] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.449078] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.449083] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa89c000b90 00:29:05.779 [2024-10-17 17:50:13.449093] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.449471] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1e76ff0 is same with the state(6) to be set 00:29:05.779 [2024-10-17 17:50:13.459052] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.459173] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.459238] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.459264] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.459285] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa8a4000b90 00:29:05.779 [2024-10-17 17:50:13.459339] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.469061] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.469127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.469159] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.469175] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.469189] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa8a4000b90 00:29:05.779 [2024-10-17 17:50:13.469221] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.479089] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.479192] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.479257] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.479283] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.479304] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa898000b90 00:29:05.779 [2024-10-17 17:50:13.479356] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.489206] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.489275] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.489307] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.489323] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.489337] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fa898000b90 00:29:05.779 [2024-10-17 17:50:13.489368] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.499159] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.499264] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.499328] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.499354] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.499374] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1e690d0 00:29:05.779 [2024-10-17 17:50:13.499424] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.509196] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:29:05.779 [2024-10-17 17:50:13.509264] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:29:05.779 [2024-10-17 17:50:13.509301] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:29:05.779 [2024-10-17 17:50:13.509318] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:29:05.779 [2024-10-17 17:50:13.509332] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1e690d0 00:29:05.779 [2024-10-17 17:50:13.509360] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:29:05.779 qpair failed and we were unable to recover it. 00:29:05.779 [2024-10-17 17:50:13.509978] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1e76ff0 (9): Bad file descriptor 00:29:05.779 Initializing NVMe Controllers 00:29:05.779 Attaching to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:29:05.779 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:29:05.779 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 0 00:29:05.779 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 1 00:29:05.779 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 2 00:29:05.779 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 3 00:29:05.779 Initialization complete. Launching workers. 00:29:05.779 Starting thread on core 1 00:29:05.779 Starting thread on core 2 00:29:05.779 Starting thread on core 3 00:29:05.779 Starting thread on core 0 00:29:05.779 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@51 -- # sync 00:29:05.779 00:29:05.779 real 0m11.450s 00:29:05.779 user 0m22.245s 00:29:05.779 sys 0m3.809s 00:29:05.779 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:05.779 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:29:05.779 ************************************ 00:29:05.779 END TEST nvmf_target_disconnect_tc2 00:29:05.779 ************************************ 00:29:05.779 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@72 -- # '[' -n '' ']' 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@76 -- # trap - SIGINT SIGTERM EXIT 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@77 -- # nvmftestfini 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@514 -- # nvmfcleanup 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@121 -- # sync 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@124 -- # set +e 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:29:05.780 rmmod nvme_tcp 00:29:05.780 rmmod nvme_fabrics 00:29:05.780 rmmod nvme_keyring 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@128 -- # set -e 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@129 -- # return 0 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@515 -- # '[' -n 234876 ']' 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@516 -- # killprocess 234876 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@950 -- # '[' -z 234876 ']' 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@954 -- # kill -0 234876 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@955 -- # uname 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:05.780 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 234876 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@956 -- # process_name=reactor_4 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@960 -- # '[' reactor_4 = sudo ']' 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@968 -- # echo 'killing process with pid 234876' 00:29:06.040 killing process with pid 234876 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@969 -- # kill 234876 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@974 -- # wait 234876 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@297 -- # iptr 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@789 -- # iptables-save 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@789 -- # iptables-restore 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:06.040 17:50:13 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:08.589 17:50:15 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:29:08.589 00:29:08.589 real 0m21.951s 00:29:08.589 user 0m50.307s 00:29:08.589 sys 0m10.059s 00:29:08.589 17:50:15 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:08.589 17:50:15 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:29:08.589 ************************************ 00:29:08.589 END TEST nvmf_target_disconnect 00:29:08.589 ************************************ 00:29:08.589 17:50:15 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:29:08.589 00:29:08.589 real 6m34.925s 00:29:08.589 user 11m20.664s 00:29:08.589 sys 2m17.031s 00:29:08.589 17:50:15 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:08.589 17:50:15 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:29:08.589 ************************************ 00:29:08.589 END TEST nvmf_host 00:29:08.589 ************************************ 00:29:08.589 17:50:16 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ tcp = \t\c\p ]] 00:29:08.589 17:50:16 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ 0 -eq 0 ]] 00:29:08.589 17:50:16 nvmf_tcp -- nvmf/nvmf.sh@20 -- # run_test nvmf_target_core_interrupt_mode /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:29:08.589 17:50:16 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:29:08.589 17:50:16 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:08.589 17:50:16 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:29:08.589 ************************************ 00:29:08.589 START TEST nvmf_target_core_interrupt_mode 00:29:08.589 ************************************ 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:29:08.589 * Looking for test storage... 00:29:08.589 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1691 -- # lcov --version 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@333 -- # local ver1 ver1_l 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@334 -- # local ver2 ver2_l 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # IFS=.-: 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # read -ra ver1 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # IFS=.-: 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # read -ra ver2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@338 -- # local 'op=<' 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@340 -- # ver1_l=2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@341 -- # ver2_l=1 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@344 -- # case "$op" in 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@345 -- # : 1 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v = 0 )) 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # decimal 1 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=1 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 1 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # ver1[v]=1 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # decimal 2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # ver2[v]=2 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # return 0 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:29:08.589 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.589 --rc genhtml_branch_coverage=1 00:29:08.589 --rc genhtml_function_coverage=1 00:29:08.589 --rc genhtml_legend=1 00:29:08.589 --rc geninfo_all_blocks=1 00:29:08.589 --rc geninfo_unexecuted_blocks=1 00:29:08.589 00:29:08.589 ' 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:29:08.589 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.589 --rc genhtml_branch_coverage=1 00:29:08.589 --rc genhtml_function_coverage=1 00:29:08.589 --rc genhtml_legend=1 00:29:08.589 --rc geninfo_all_blocks=1 00:29:08.589 --rc geninfo_unexecuted_blocks=1 00:29:08.589 00:29:08.589 ' 00:29:08.589 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:29:08.589 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.589 --rc genhtml_branch_coverage=1 00:29:08.589 --rc genhtml_function_coverage=1 00:29:08.589 --rc genhtml_legend=1 00:29:08.589 --rc geninfo_all_blocks=1 00:29:08.589 --rc geninfo_unexecuted_blocks=1 00:29:08.589 00:29:08.589 ' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:29:08.590 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.590 --rc genhtml_branch_coverage=1 00:29:08.590 --rc genhtml_function_coverage=1 00:29:08.590 --rc genhtml_legend=1 00:29:08.590 --rc geninfo_all_blocks=1 00:29:08.590 --rc geninfo_unexecuted_blocks=1 00:29:08.590 00:29:08.590 ' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # uname -s 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@15 -- # shopt -s extglob 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@5 -- # export PATH 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@51 -- # : 0 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@55 -- # have_pci_nics=0 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:29:08.590 ************************************ 00:29:08.590 START TEST nvmf_abort 00:29:08.590 ************************************ 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:29:08.590 * Looking for test storage... 00:29:08.590 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1691 -- # lcov --version 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:29:08.590 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:29:08.852 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.852 --rc genhtml_branch_coverage=1 00:29:08.852 --rc genhtml_function_coverage=1 00:29:08.852 --rc genhtml_legend=1 00:29:08.852 --rc geninfo_all_blocks=1 00:29:08.852 --rc geninfo_unexecuted_blocks=1 00:29:08.852 00:29:08.852 ' 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:29:08.852 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.852 --rc genhtml_branch_coverage=1 00:29:08.852 --rc genhtml_function_coverage=1 00:29:08.852 --rc genhtml_legend=1 00:29:08.852 --rc geninfo_all_blocks=1 00:29:08.852 --rc geninfo_unexecuted_blocks=1 00:29:08.852 00:29:08.852 ' 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:29:08.852 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.852 --rc genhtml_branch_coverage=1 00:29:08.852 --rc genhtml_function_coverage=1 00:29:08.852 --rc genhtml_legend=1 00:29:08.852 --rc geninfo_all_blocks=1 00:29:08.852 --rc geninfo_unexecuted_blocks=1 00:29:08.852 00:29:08.852 ' 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:29:08.852 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:08.852 --rc genhtml_branch_coverage=1 00:29:08.852 --rc genhtml_function_coverage=1 00:29:08.852 --rc genhtml_legend=1 00:29:08.852 --rc geninfo_all_blocks=1 00:29:08.852 --rc geninfo_unexecuted_blocks=1 00:29:08.852 00:29:08.852 ' 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:08.852 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@474 -- # prepare_net_devs 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@436 -- # local -g is_hw=no 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@438 -- # remove_spdk_ns 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:29:08.853 17:50:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:29:16.995 Found 0000:31:00.0 (0x8086 - 0x159b) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:29:16.995 Found 0000:31:00.1 (0x8086 - 0x159b) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:29:16.995 Found net devices under 0000:31:00.0: cvl_0_0 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:29:16.995 Found net devices under 0000:31:00.1: cvl_0_1 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # is_hw=yes 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:29:16.995 17:50:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:29:16.995 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:29:16.995 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:29:16.995 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:29:16.995 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:29:16.995 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:29:16.995 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.662 ms 00:29:16.995 00:29:16.995 --- 10.0.0.2 ping statistics --- 00:29:16.995 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:16.995 rtt min/avg/max/mdev = 0.662/0.662/0.662/0.000 ms 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:29:16.996 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:29:16.996 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.317 ms 00:29:16.996 00:29:16.996 --- 10.0.0.1 ping statistics --- 00:29:16.996 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:16.996 rtt min/avg/max/mdev = 0.317/0.317/0.317/0.000 ms 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@448 -- # return 0 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@724 -- # xtrace_disable 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@507 -- # nvmfpid=240359 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@508 -- # waitforlisten 240359 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@831 -- # '[' -z 240359 ']' 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:16.996 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:16.996 17:50:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:16.996 [2024-10-17 17:50:24.235672] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:29:16.996 [2024-10-17 17:50:24.236851] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:29:16.996 [2024-10-17 17:50:24.236898] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:16.996 [2024-10-17 17:50:24.329550] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:29:16.996 [2024-10-17 17:50:24.382645] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:16.996 [2024-10-17 17:50:24.382708] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:16.996 [2024-10-17 17:50:24.382718] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:16.996 [2024-10-17 17:50:24.382730] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:16.996 [2024-10-17 17:50:24.382736] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:16.996 [2024-10-17 17:50:24.384638] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:29:16.996 [2024-10-17 17:50:24.384800] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:29:16.996 [2024-10-17 17:50:24.385008] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:16.996 [2024-10-17 17:50:24.462357] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:29:16.996 [2024-10-17 17:50:24.463407] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:29:16.996 [2024-10-17 17:50:24.463621] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:29:16.996 [2024-10-17 17:50:24.463800] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@864 -- # return 0 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@730 -- # xtrace_disable 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.257 [2024-10-17 17:50:25.105947] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.257 Malloc0 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.257 Delay0 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.257 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.518 [2024-10-17 17:50:25.205892] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.518 17:50:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:29:17.518 [2024-10-17 17:50:25.342912] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:29:20.059 Initializing NVMe Controllers 00:29:20.059 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:29:20.059 controller IO queue size 128 less than required 00:29:20.059 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:29:20.059 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:29:20.059 Initialization complete. Launching workers. 00:29:20.059 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 127, failed: 28594 00:29:20.059 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 28655, failed to submit 66 00:29:20.059 success 28594, unsuccessful 61, failed 0 00:29:20.059 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:29:20.059 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:20.059 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@514 -- # nvmfcleanup 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:29:20.060 rmmod nvme_tcp 00:29:20.060 rmmod nvme_fabrics 00:29:20.060 rmmod nvme_keyring 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@515 -- # '[' -n 240359 ']' 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@516 -- # killprocess 240359 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@950 -- # '[' -z 240359 ']' 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@954 -- # kill -0 240359 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@955 -- # uname 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 240359 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@968 -- # echo 'killing process with pid 240359' 00:29:20.060 killing process with pid 240359 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@969 -- # kill 240359 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@974 -- # wait 240359 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@789 -- # iptables-save 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@789 -- # iptables-restore 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:20.060 17:50:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:29:21.972 00:29:21.972 real 0m13.485s 00:29:21.972 user 0m10.809s 00:29:21.972 sys 0m7.001s 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:29:21.972 ************************************ 00:29:21.972 END TEST nvmf_abort 00:29:21.972 ************************************ 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:29:21.972 ************************************ 00:29:21.972 START TEST nvmf_ns_hotplug_stress 00:29:21.972 ************************************ 00:29:21.972 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:29:22.234 * Looking for test storage... 00:29:22.234 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:29:22.234 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:29:22.234 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1691 -- # lcov --version 00:29:22.234 17:50:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:29:22.234 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:29:22.235 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.235 --rc genhtml_branch_coverage=1 00:29:22.235 --rc genhtml_function_coverage=1 00:29:22.235 --rc genhtml_legend=1 00:29:22.235 --rc geninfo_all_blocks=1 00:29:22.235 --rc geninfo_unexecuted_blocks=1 00:29:22.235 00:29:22.235 ' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:29:22.235 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.235 --rc genhtml_branch_coverage=1 00:29:22.235 --rc genhtml_function_coverage=1 00:29:22.235 --rc genhtml_legend=1 00:29:22.235 --rc geninfo_all_blocks=1 00:29:22.235 --rc geninfo_unexecuted_blocks=1 00:29:22.235 00:29:22.235 ' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:29:22.235 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.235 --rc genhtml_branch_coverage=1 00:29:22.235 --rc genhtml_function_coverage=1 00:29:22.235 --rc genhtml_legend=1 00:29:22.235 --rc geninfo_all_blocks=1 00:29:22.235 --rc geninfo_unexecuted_blocks=1 00:29:22.235 00:29:22.235 ' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:29:22.235 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:22.235 --rc genhtml_branch_coverage=1 00:29:22.235 --rc genhtml_function_coverage=1 00:29:22.235 --rc genhtml_legend=1 00:29:22.235 --rc geninfo_all_blocks=1 00:29:22.235 --rc geninfo_unexecuted_blocks=1 00:29:22.235 00:29:22.235 ' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # prepare_net_devs 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@436 -- # local -g is_hw=no 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # remove_spdk_ns 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:22.235 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:29:22.236 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:29:22.236 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:29:22.236 17:50:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:29:30.377 Found 0000:31:00.0 (0x8086 - 0x159b) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:29:30.377 Found 0000:31:00.1 (0x8086 - 0x159b) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:29:30.377 Found net devices under 0000:31:00.0: cvl_0_0 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:29:30.377 Found net devices under 0000:31:00.1: cvl_0_1 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:30.377 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # is_hw=yes 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:29:30.378 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:29:30.378 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.602 ms 00:29:30.378 00:29:30.378 --- 10.0.0.2 ping statistics --- 00:29:30.378 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:30.378 rtt min/avg/max/mdev = 0.602/0.602/0.602/0.000 ms 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:29:30.378 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:29:30.378 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.299 ms 00:29:30.378 00:29:30.378 --- 10.0.0.1 ping statistics --- 00:29:30.378 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:30.378 rtt min/avg/max/mdev = 0.299/0.299/0.299/0.000 ms 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@448 -- # return 0 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@724 -- # xtrace_disable 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # nvmfpid=245342 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # waitforlisten 245342 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@831 -- # '[' -z 245342 ']' 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:30.378 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:30.378 17:50:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:29:30.378 [2024-10-17 17:50:37.779373] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:29:30.378 [2024-10-17 17:50:37.780379] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:29:30.378 [2024-10-17 17:50:37.780414] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:30.378 [2024-10-17 17:50:37.864143] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:29:30.378 [2024-10-17 17:50:37.900229] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:30.378 [2024-10-17 17:50:37.900261] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:30.378 [2024-10-17 17:50:37.900269] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:30.378 [2024-10-17 17:50:37.900276] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:30.378 [2024-10-17 17:50:37.900282] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:30.378 [2024-10-17 17:50:37.901621] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:29:30.378 [2024-10-17 17:50:37.901748] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:29:30.378 [2024-10-17 17:50:37.901931] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:30.378 [2024-10-17 17:50:37.956651] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:29:30.378 [2024-10-17 17:50:37.957629] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:29:30.378 [2024-10-17 17:50:37.958678] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:29:30.378 [2024-10-17 17:50:37.958776] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # return 0 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@730 -- # xtrace_disable 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:29:30.968 [2024-10-17 17:50:38.779005] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:30.968 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:29:31.277 17:50:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:29:31.277 [2024-10-17 17:50:39.163645] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:31.277 17:50:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:29:31.549 17:50:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:29:31.820 Malloc0 00:29:31.820 17:50:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:29:31.820 Delay0 00:29:31.820 17:50:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:32.080 17:50:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:29:32.340 NULL1 00:29:32.340 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:29:32.601 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=245761 00:29:32.601 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:32.601 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:29:32.601 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:32.601 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:32.863 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:29:32.863 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:29:33.125 true 00:29:33.125 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:33.125 17:50:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:33.387 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:33.387 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:29:33.387 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:29:33.649 true 00:29:33.649 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:33.649 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:33.909 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:34.169 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:29:34.169 17:50:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:29:34.169 true 00:29:34.169 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:34.169 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:34.429 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:34.690 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:29:34.690 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:29:34.690 true 00:29:34.951 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:34.951 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:34.951 17:50:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:35.212 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:29:35.212 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:29:35.473 true 00:29:35.473 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:35.473 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:35.473 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:35.734 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:29:35.734 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:29:35.993 true 00:29:35.993 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:35.993 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:36.253 17:50:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:36.253 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:29:36.253 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:29:36.512 true 00:29:36.512 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:36.512 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:36.772 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:36.772 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:29:36.772 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:29:37.033 true 00:29:37.033 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:37.033 17:50:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:37.294 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:37.555 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:29:37.555 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:29:37.555 true 00:29:37.555 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:37.555 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:37.815 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:38.076 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:29:38.076 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:29:38.076 true 00:29:38.076 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:38.076 17:50:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:38.336 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:38.598 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:29:38.598 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:29:38.598 true 00:29:38.598 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:38.598 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:38.859 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:39.120 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:29:39.120 17:50:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:29:39.120 true 00:29:39.381 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:39.381 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:39.381 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:39.642 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:29:39.642 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:29:39.903 true 00:29:39.903 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:39.903 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:39.903 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:40.165 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:29:40.165 17:50:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:29:40.426 true 00:29:40.426 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:40.426 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:40.426 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:40.687 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:29:40.687 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:29:40.947 true 00:29:40.947 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:40.947 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:41.207 17:50:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:41.208 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:29:41.208 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:29:41.468 true 00:29:41.468 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:41.468 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:41.729 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:41.729 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:29:41.729 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:29:41.990 true 00:29:41.990 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:41.990 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:42.251 17:50:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:42.511 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:29:42.511 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:29:42.511 true 00:29:42.511 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:42.511 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:42.771 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:43.033 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:29:43.033 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:29:43.033 true 00:29:43.033 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:43.033 17:50:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:43.293 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:43.554 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:29:43.554 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:29:43.554 true 00:29:43.554 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:43.554 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:43.814 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:44.075 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:29:44.075 17:50:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:29:44.075 true 00:29:44.335 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:44.335 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:44.335 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:44.595 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:29:44.595 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:29:44.855 true 00:29:44.855 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:44.855 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:44.855 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:45.116 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:29:45.116 17:50:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:29:45.377 true 00:29:45.377 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:45.377 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:45.638 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:45.638 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:29:45.639 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:29:45.900 true 00:29:45.900 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:45.900 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:46.162 17:50:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:46.162 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:29:46.162 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:29:46.423 true 00:29:46.423 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:46.423 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:46.684 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:46.684 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:29:46.684 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:29:46.945 true 00:29:46.945 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:46.945 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:47.205 17:50:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:47.466 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:29:47.466 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:29:47.466 true 00:29:47.466 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:47.466 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:47.727 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:47.988 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:29:47.988 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:29:47.988 true 00:29:47.988 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:47.988 17:50:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:48.249 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:48.510 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:29:48.510 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:29:48.510 true 00:29:48.771 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:48.771 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:48.771 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:49.032 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1030 00:29:49.032 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1030 00:29:49.293 true 00:29:49.293 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:49.293 17:50:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:49.293 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:49.554 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1031 00:29:49.554 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1031 00:29:49.815 true 00:29:49.815 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:49.815 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:49.815 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:50.076 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1032 00:29:50.076 17:50:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1032 00:29:50.338 true 00:29:50.338 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:50.338 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:50.600 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:50.600 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1033 00:29:50.600 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1033 00:29:50.861 true 00:29:50.861 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:50.861 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:51.123 17:50:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:51.123 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1034 00:29:51.123 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1034 00:29:51.385 true 00:29:51.385 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:51.385 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:51.645 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:51.907 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1035 00:29:51.907 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1035 00:29:51.907 true 00:29:51.907 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:51.907 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:52.168 17:50:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:52.429 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1036 00:29:52.429 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1036 00:29:52.429 true 00:29:52.690 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:52.690 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:52.690 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:52.951 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1037 00:29:52.951 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1037 00:29:52.951 true 00:29:53.213 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:53.213 17:51:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:53.213 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:53.474 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1038 00:29:53.474 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1038 00:29:53.735 true 00:29:53.735 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:53.735 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:53.735 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:53.996 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1039 00:29:53.996 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1039 00:29:54.257 true 00:29:54.257 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:54.257 17:51:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:54.257 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:54.518 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1040 00:29:54.518 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1040 00:29:54.779 true 00:29:54.779 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:54.779 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:55.040 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:55.040 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1041 00:29:55.040 17:51:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1041 00:29:55.300 true 00:29:55.300 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:55.300 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:55.560 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:55.560 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1042 00:29:55.560 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1042 00:29:55.822 true 00:29:55.822 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:55.822 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:56.083 17:51:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:56.344 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1043 00:29:56.344 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1043 00:29:56.344 true 00:29:56.344 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:56.344 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:56.605 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:56.866 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1044 00:29:56.866 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1044 00:29:56.866 true 00:29:56.866 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:56.866 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:57.127 17:51:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:57.390 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1045 00:29:57.390 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1045 00:29:57.390 true 00:29:57.651 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:57.651 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:57.651 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:57.911 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1046 00:29:57.911 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1046 00:29:58.172 true 00:29:58.172 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:58.172 17:51:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:58.172 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:58.433 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1047 00:29:58.433 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1047 00:29:58.694 true 00:29:58.694 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:58.694 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:58.694 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:58.954 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1048 00:29:58.954 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1048 00:29:59.214 true 00:29:59.214 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:59.214 17:51:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:59.474 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:59.475 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1049 00:29:59.475 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1049 00:29:59.736 true 00:29:59.736 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:29:59.736 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:29:59.997 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:00.257 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1050 00:30:00.257 17:51:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1050 00:30:00.257 true 00:30:00.257 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:30:00.257 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:00.517 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:00.778 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1051 00:30:00.779 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1051 00:30:00.779 true 00:30:00.779 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:30:00.779 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:01.039 17:51:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:01.300 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1052 00:30:01.300 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1052 00:30:01.300 true 00:30:01.300 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:30:01.300 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:01.561 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:01.821 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1053 00:30:01.821 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1053 00:30:01.821 true 00:30:02.082 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:30:02.082 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:02.082 17:51:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:02.344 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1054 00:30:02.344 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1054 00:30:02.605 true 00:30:02.605 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:30:02.605 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:02.605 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:02.605 Initializing NVMe Controllers 00:30:02.605 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:30:02.605 Controller SPDK bdev Controller (SPDK00000000000001 ): Skipping inactive NS 1 00:30:02.605 Controller IO queue size 128, less than required. 00:30:02.605 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:30:02.605 WARNING: Some requested NVMe devices were skipped 00:30:02.605 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:30:02.605 Initialization complete. Launching workers. 00:30:02.605 ======================================================== 00:30:02.605 Latency(us) 00:30:02.605 Device Information : IOPS MiB/s Average min max 00:30:02.605 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 30607.90 14.95 4181.88 1136.67 10901.28 00:30:02.605 ======================================================== 00:30:02.605 Total : 30607.90 14.95 4181.88 1136.67 10901.28 00:30:02.605 00:30:02.866 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1055 00:30:02.866 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1055 00:30:03.127 true 00:30:03.127 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 245761 00:30:03.127 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (245761) - No such process 00:30:03.127 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 245761 00:30:03.127 17:51:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:03.388 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:03.388 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:30:03.388 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:30:03.388 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:30:03.388 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:03.388 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:30:03.649 null0 00:30:03.649 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:03.649 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:03.649 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:30:03.649 null1 00:30:03.909 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:03.909 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:03.909 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:30:03.909 null2 00:30:03.909 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:03.909 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:03.909 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:30:04.170 null3 00:30:04.170 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:04.170 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:04.170 17:51:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:30:04.431 null4 00:30:04.431 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:04.431 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:04.431 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:30:04.431 null5 00:30:04.431 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:04.431 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:04.431 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:30:04.691 null6 00:30:04.691 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:04.691 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:04.691 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:30:04.953 null7 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:30:04.953 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 252564 252566 252569 252572 252575 252578 252580 252582 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:04.954 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:05.215 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:05.215 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:05.215 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:05.215 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:05.216 17:51:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.216 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:05.477 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.738 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:05.739 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.000 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:06.260 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:06.260 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:06.260 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:06.260 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:06.260 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:06.260 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:06.260 17:51:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.260 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:06.521 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:06.781 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:06.781 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:06.782 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.043 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:07.305 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.305 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.305 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:07.305 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.305 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.305 17:51:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.305 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:07.567 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:07.829 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.091 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.092 17:51:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:30:08.353 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@514 -- # nvmfcleanup 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:08.615 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:08.615 rmmod nvme_tcp 00:30:08.615 rmmod nvme_fabrics 00:30:08.876 rmmod nvme_keyring 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@515 -- # '[' -n 245342 ']' 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # killprocess 245342 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@950 -- # '[' -z 245342 ']' 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # kill -0 245342 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # uname 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 245342 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@968 -- # echo 'killing process with pid 245342' 00:30:08.876 killing process with pid 245342 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@969 -- # kill 245342 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@974 -- # wait 245342 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-save 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:30:08.876 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-restore 00:30:09.138 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:09.138 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:09.138 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:09.138 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:09.138 17:51:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:11.054 00:30:11.054 real 0m48.989s 00:30:11.054 user 3m1.122s 00:30:11.054 sys 0m23.235s 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:30:11.054 ************************************ 00:30:11.054 END TEST nvmf_ns_hotplug_stress 00:30:11.054 ************************************ 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:30:11.054 ************************************ 00:30:11.054 START TEST nvmf_delete_subsystem 00:30:11.054 ************************************ 00:30:11.054 17:51:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:30:11.348 * Looking for test storage... 00:30:11.348 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1691 -- # lcov --version 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:30:11.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:11.348 --rc genhtml_branch_coverage=1 00:30:11.348 --rc genhtml_function_coverage=1 00:30:11.348 --rc genhtml_legend=1 00:30:11.348 --rc geninfo_all_blocks=1 00:30:11.348 --rc geninfo_unexecuted_blocks=1 00:30:11.348 00:30:11.348 ' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:30:11.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:11.348 --rc genhtml_branch_coverage=1 00:30:11.348 --rc genhtml_function_coverage=1 00:30:11.348 --rc genhtml_legend=1 00:30:11.348 --rc geninfo_all_blocks=1 00:30:11.348 --rc geninfo_unexecuted_blocks=1 00:30:11.348 00:30:11.348 ' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:30:11.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:11.348 --rc genhtml_branch_coverage=1 00:30:11.348 --rc genhtml_function_coverage=1 00:30:11.348 --rc genhtml_legend=1 00:30:11.348 --rc geninfo_all_blocks=1 00:30:11.348 --rc geninfo_unexecuted_blocks=1 00:30:11.348 00:30:11.348 ' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:30:11.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:11.348 --rc genhtml_branch_coverage=1 00:30:11.348 --rc genhtml_function_coverage=1 00:30:11.348 --rc genhtml_legend=1 00:30:11.348 --rc geninfo_all_blocks=1 00:30:11.348 --rc geninfo_unexecuted_blocks=1 00:30:11.348 00:30:11.348 ' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:11.348 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # prepare_net_devs 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@436 -- # local -g is_hw=no 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # remove_spdk_ns 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:30:11.349 17:51:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:30:19.564 Found 0000:31:00.0 (0x8086 - 0x159b) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:30:19.564 Found 0000:31:00.1 (0x8086 - 0x159b) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:30:19.564 Found net devices under 0000:31:00.0: cvl_0_0 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:30:19.564 Found net devices under 0000:31:00.1: cvl_0_1 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # is_hw=yes 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:19.564 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:19.565 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:19.565 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.641 ms 00:30:19.565 00:30:19.565 --- 10.0.0.2 ping statistics --- 00:30:19.565 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:19.565 rtt min/avg/max/mdev = 0.641/0.641/0.641/0.000 ms 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:19.565 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:19.565 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.279 ms 00:30:19.565 00:30:19.565 --- 10.0.0.1 ping statistics --- 00:30:19.565 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:19.565 rtt min/avg/max/mdev = 0.279/0.279/0.279/0.000 ms 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@448 -- # return 0 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@724 -- # xtrace_disable 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # nvmfpid=257703 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # waitforlisten 257703 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@831 -- # '[' -z 257703 ']' 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:19.565 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:19.565 17:51:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:19.565 [2024-10-17 17:51:26.853775] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:30:19.565 [2024-10-17 17:51:26.854936] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:30:19.565 [2024-10-17 17:51:26.854983] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:19.565 [2024-10-17 17:51:26.944822] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:30:19.565 [2024-10-17 17:51:26.996510] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:19.565 [2024-10-17 17:51:26.996556] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:19.565 [2024-10-17 17:51:26.996565] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:19.565 [2024-10-17 17:51:26.996572] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:19.565 [2024-10-17 17:51:26.996579] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:19.565 [2024-10-17 17:51:26.998304] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:30:19.565 [2024-10-17 17:51:26.998308] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:19.565 [2024-10-17 17:51:27.074591] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:30:19.565 [2024-10-17 17:51:27.075227] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:30:19.565 [2024-10-17 17:51:27.075503] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # return 0 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@730 -- # xtrace_disable 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.825 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:19.825 [2024-10-17 17:51:27.731321] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:20.086 [2024-10-17 17:51:27.763845] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:20.086 NULL1 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:20.086 Delay0 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=258046 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:30:20.086 17:51:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:30:20.086 [2024-10-17 17:51:27.874841] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:30:21.997 17:51:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:30:21.997 17:51:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.997 17:51:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 starting I/O failed: -6 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 [2024-10-17 17:51:29.958155] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xaea1b0 is same with the state(6) to be set 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Write completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.258 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 starting I/O failed: -6 00:30:22.259 [2024-10-17 17:51:29.962380] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f994400d450 is same with the state(6) to be set 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Read completed with error (sct=0, sc=8) 00:30:22.259 Write completed with error (sct=0, sc=8) 00:30:23.201 [2024-10-17 17:51:30.934808] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xaeb6b0 is same with the state(6) to be set 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 [2024-10-17 17:51:30.961479] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xaea390 is same with the state(6) to be set 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 [2024-10-17 17:51:30.962043] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xae9fd0 is same with the state(6) to be set 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 [2024-10-17 17:51:30.964540] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f994400d780 is same with the state(6) to be set 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Read completed with error (sct=0, sc=8) 00:30:23.201 Write completed with error (sct=0, sc=8) 00:30:23.201 [2024-10-17 17:51:30.964635] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f994400cfe0 is same with the state(6) to be set 00:30:23.201 Initializing NVMe Controllers 00:30:23.201 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:30:23.201 Controller IO queue size 128, less than required. 00:30:23.201 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:30:23.201 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:30:23.201 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:30:23.201 Initialization complete. Launching workers. 00:30:23.201 ======================================================== 00:30:23.201 Latency(us) 00:30:23.201 Device Information : IOPS MiB/s Average min max 00:30:23.201 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 170.14 0.08 895694.60 422.80 1008217.42 00:30:23.202 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 165.17 0.08 905745.18 327.87 1011939.00 00:30:23.202 ======================================================== 00:30:23.202 Total : 335.31 0.16 900645.33 327.87 1011939.00 00:30:23.202 00:30:23.202 [2024-10-17 17:51:30.965239] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xaeb6b0 (9): Bad file descriptor 00:30:23.202 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:30:23.202 17:51:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.202 17:51:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:30:23.202 17:51:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 258046 00:30:23.202 17:51:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 258046 00:30:23.774 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (258046) - No such process 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 258046 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@650 -- # local es=0 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # valid_exec_arg wait 258046 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@638 -- # local arg=wait 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # type -t wait 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # wait 258046 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # es=1 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:23.774 [2024-10-17 17:51:31.499614] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=258715 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:30:23.774 17:51:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:30:23.774 [2024-10-17 17:51:31.584817] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:30:24.345 17:51:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:30:24.345 17:51:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:24.345 17:51:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:30:24.915 17:51:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:30:24.915 17:51:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:24.915 17:51:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:30:25.175 17:51:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:30:25.175 17:51:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:25.175 17:51:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:30:25.744 17:51:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:30:25.744 17:51:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:25.744 17:51:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:30:26.315 17:51:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:30:26.315 17:51:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:26.315 17:51:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:30:26.885 17:51:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:30:26.885 17:51:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:26.885 17:51:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:30:26.885 Initializing NVMe Controllers 00:30:26.885 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:30:26.885 Controller IO queue size 128, less than required. 00:30:26.885 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:30:26.885 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:30:26.885 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:30:26.885 Initialization complete. Launching workers. 00:30:26.885 ======================================================== 00:30:26.885 Latency(us) 00:30:26.885 Device Information : IOPS MiB/s Average min max 00:30:26.885 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1002399.18 1000283.84 1041839.41 00:30:26.885 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1004054.93 1000278.87 1010459.93 00:30:26.885 ======================================================== 00:30:26.886 Total : 256.00 0.12 1003227.06 1000278.87 1041839.41 00:30:26.886 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 258715 00:30:27.146 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (258715) - No such process 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 258715 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@514 -- # nvmfcleanup 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:27.146 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:27.146 rmmod nvme_tcp 00:30:27.406 rmmod nvme_fabrics 00:30:27.406 rmmod nvme_keyring 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@515 -- # '[' -n 257703 ']' 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # killprocess 257703 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@950 -- # '[' -z 257703 ']' 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # kill -0 257703 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # uname 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 257703 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@968 -- # echo 'killing process with pid 257703' 00:30:27.406 killing process with pid 257703 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@969 -- # kill 257703 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@974 -- # wait 257703 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:30:27.406 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-save 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-restore 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:27.407 17:51:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:29.951 00:30:29.951 real 0m18.410s 00:30:29.951 user 0m26.382s 00:30:29.951 sys 0m7.534s 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:30:29.951 ************************************ 00:30:29.951 END TEST nvmf_delete_subsystem 00:30:29.951 ************************************ 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:30:29.951 ************************************ 00:30:29.951 START TEST nvmf_host_management 00:30:29.951 ************************************ 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:30:29.951 * Looking for test storage... 00:30:29.951 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1691 -- # lcov --version 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:29.951 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:30:29.952 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:29.952 --rc genhtml_branch_coverage=1 00:30:29.952 --rc genhtml_function_coverage=1 00:30:29.952 --rc genhtml_legend=1 00:30:29.952 --rc geninfo_all_blocks=1 00:30:29.952 --rc geninfo_unexecuted_blocks=1 00:30:29.952 00:30:29.952 ' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:30:29.952 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:29.952 --rc genhtml_branch_coverage=1 00:30:29.952 --rc genhtml_function_coverage=1 00:30:29.952 --rc genhtml_legend=1 00:30:29.952 --rc geninfo_all_blocks=1 00:30:29.952 --rc geninfo_unexecuted_blocks=1 00:30:29.952 00:30:29.952 ' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:30:29.952 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:29.952 --rc genhtml_branch_coverage=1 00:30:29.952 --rc genhtml_function_coverage=1 00:30:29.952 --rc genhtml_legend=1 00:30:29.952 --rc geninfo_all_blocks=1 00:30:29.952 --rc geninfo_unexecuted_blocks=1 00:30:29.952 00:30:29.952 ' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:30:29.952 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:29.952 --rc genhtml_branch_coverage=1 00:30:29.952 --rc genhtml_function_coverage=1 00:30:29.952 --rc genhtml_legend=1 00:30:29.952 --rc geninfo_all_blocks=1 00:30:29.952 --rc geninfo_unexecuted_blocks=1 00:30:29.952 00:30:29.952 ' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:29.952 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@474 -- # prepare_net_devs 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@436 -- # local -g is_hw=no 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@438 -- # remove_spdk_ns 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:30:29.953 17:51:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:30:38.099 Found 0000:31:00.0 (0x8086 - 0x159b) 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:38.099 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:30:38.100 Found 0000:31:00.1 (0x8086 - 0x159b) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:30:38.100 Found net devices under 0000:31:00.0: cvl_0_0 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:30:38.100 Found net devices under 0000:31:00.1: cvl_0_1 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # is_hw=yes 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:38.100 17:51:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:38.100 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:38.100 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.465 ms 00:30:38.100 00:30:38.100 --- 10.0.0.2 ping statistics --- 00:30:38.100 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:38.100 rtt min/avg/max/mdev = 0.465/0.465/0.465/0.000 ms 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:38.100 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:38.100 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.180 ms 00:30:38.100 00:30:38.100 --- 10.0.0.1 ping statistics --- 00:30:38.100 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:38.100 rtt min/avg/max/mdev = 0.180/0.180/0.180/0.000 ms 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@448 -- # return 0 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@507 -- # nvmfpid=263514 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@508 -- # waitforlisten 263514 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1E 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 263514 ']' 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:38.100 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:38.100 17:51:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.100 [2024-10-17 17:51:45.381577] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:30:38.100 [2024-10-17 17:51:45.382780] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:30:38.100 [2024-10-17 17:51:45.382833] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:38.100 [2024-10-17 17:51:45.473814] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:30:38.100 [2024-10-17 17:51:45.527274] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:38.100 [2024-10-17 17:51:45.527326] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:38.100 [2024-10-17 17:51:45.527335] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:38.100 [2024-10-17 17:51:45.527343] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:38.100 [2024-10-17 17:51:45.527349] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:38.100 [2024-10-17 17:51:45.529435] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:30:38.100 [2024-10-17 17:51:45.529591] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:30:38.101 [2024-10-17 17:51:45.529752] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:30:38.101 [2024-10-17 17:51:45.529752] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:30:38.101 [2024-10-17 17:51:45.606946] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:30:38.101 [2024-10-17 17:51:45.607829] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:30:38.101 [2024-10-17 17:51:45.608258] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:30:38.101 [2024-10-17 17:51:45.608763] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:30:38.101 [2024-10-17 17:51:45.608800] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:30:38.361 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:38.361 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:30:38.361 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:30:38.361 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:30:38.361 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.361 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.362 [2024-10-17 17:51:46.242604] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:30:38.362 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.626 Malloc0 00:30:38.626 [2024-10-17 17:51:46.338933] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=263801 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 263801 /var/tmp/bdevperf.sock 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 263801 ']' 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:30:38.626 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:30:38.626 { 00:30:38.626 "params": { 00:30:38.626 "name": "Nvme$subsystem", 00:30:38.626 "trtype": "$TEST_TRANSPORT", 00:30:38.626 "traddr": "$NVMF_FIRST_TARGET_IP", 00:30:38.626 "adrfam": "ipv4", 00:30:38.626 "trsvcid": "$NVMF_PORT", 00:30:38.626 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:30:38.626 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:30:38.626 "hdgst": ${hdgst:-false}, 00:30:38.626 "ddgst": ${ddgst:-false} 00:30:38.626 }, 00:30:38.626 "method": "bdev_nvme_attach_controller" 00:30:38.626 } 00:30:38.626 EOF 00:30:38.626 )") 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:30:38.626 17:51:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:30:38.626 "params": { 00:30:38.626 "name": "Nvme0", 00:30:38.626 "trtype": "tcp", 00:30:38.626 "traddr": "10.0.0.2", 00:30:38.626 "adrfam": "ipv4", 00:30:38.626 "trsvcid": "4420", 00:30:38.626 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:30:38.626 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:30:38.626 "hdgst": false, 00:30:38.626 "ddgst": false 00:30:38.626 }, 00:30:38.626 "method": "bdev_nvme_attach_controller" 00:30:38.626 }' 00:30:38.626 [2024-10-17 17:51:46.459835] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:30:38.626 [2024-10-17 17:51:46.459910] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid263801 ] 00:30:38.626 [2024-10-17 17:51:46.542014] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:38.887 [2024-10-17 17:51:46.594801] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:38.887 Running I/O for 10 seconds... 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=707 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@58 -- # '[' 707 -ge 100 ']' 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@60 -- # break 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.461 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:39.461 [2024-10-17 17:51:47.354320] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xec3fe0 is same with the state(6) to be set 00:30:39.461 [2024-10-17 17:51:47.354379] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xec3fe0 is same with the state(6) to be set 00:30:39.461 [2024-10-17 17:51:47.354389] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xec3fe0 is same with the state(6) to be set 00:30:39.461 [2024-10-17 17:51:47.354598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:106112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.461 [2024-10-17 17:51:47.354663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.461 [2024-10-17 17:51:47.354686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:106240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.461 [2024-10-17 17:51:47.354705] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.461 [2024-10-17 17:51:47.354716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:106368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.461 [2024-10-17 17:51:47.354725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.461 [2024-10-17 17:51:47.354736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:98304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.461 [2024-10-17 17:51:47.354744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:98432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:98560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:98688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:98816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:98944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:99072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:99200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:99328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:99456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:99584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:99712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:99840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.354983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:99968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.354991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:100096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:100224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:100352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:100480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:100608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:100736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:100864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:100992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:101120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:101248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:101376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:101504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355226] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:101632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:101760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:101888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:102016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:102144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:102272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:102400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:102528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:102656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:102784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:102912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:103040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:103168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:103296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:103424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.462 [2024-10-17 17:51:47.355498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.462 [2024-10-17 17:51:47.355508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:103552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:103680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:103808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:103936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:104064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:104192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:104320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:104448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:104576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:104704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:104832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:104960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:105088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:105216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:105344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:105472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:105600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:105728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:105856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:105984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:30:39.463 [2024-10-17 17:51:47.355888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.355897] nvme_tcp.c: 337:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1152570 is same with the state(6) to be set 00:30:39.463 [2024-10-17 17:51:47.355968] bdev_nvme.c:1731:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1152570 was disconnected and freed. reset controller. 00:30:39.463 [2024-10-17 17:51:47.357216] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:30:39.463 task offset: 106112 on job bdev=Nvme0n1 fails 00:30:39.463 00:30:39.463 Latency(us) 00:30:39.463 [2024-10-17T15:51:47.382Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:39.463 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:30:39.463 Job: Nvme0n1 ended in about 0.55 seconds with error 00:30:39.463 Verification LBA range: start 0x0 length 0x400 00:30:39.463 Nvme0n1 : 0.55 1385.68 86.61 115.47 0.00 41600.66 1774.93 38884.69 00:30:39.463 [2024-10-17T15:51:47.382Z] =================================================================================================================== 00:30:39.463 [2024-10-17T15:51:47.382Z] Total : 1385.68 86.61 115.47 0.00 41600.66 1774.93 38884.69 00:30:39.463 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.463 [2024-10-17 17:51:47.359469] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:30:39.463 [2024-10-17 17:51:47.359513] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1141c40 (9): Bad file descriptor 00:30:39.463 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:30:39.463 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.463 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:39.463 [2024-10-17 17:51:47.360835] ctrlr.c: 823:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode0' does not allow host 'nqn.2016-06.io.spdk:host0' 00:30:39.463 [2024-10-17 17:51:47.360932] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:3 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:30:39.463 [2024-10-17 17:51:47.360961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND SPECIFIC (01/84) qid:0 cid:3 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:39.463 [2024-10-17 17:51:47.360978] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode0 00:30:39.463 [2024-10-17 17:51:47.360987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 132 00:30:39.463 [2024-10-17 17:51:47.360996] nvme_tcp.c:2459:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:30:39.463 [2024-10-17 17:51:47.361004] nvme_tcp.c:2236:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1141c40 00:30:39.463 [2024-10-17 17:51:47.361027] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1141c40 (9): Bad file descriptor 00:30:39.463 [2024-10-17 17:51:47.361040] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:30:39.463 [2024-10-17 17:51:47.361048] nvme_ctrlr.c:1822:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:30:39.463 [2024-10-17 17:51:47.361058] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:30:39.463 [2024-10-17 17:51:47.361072] bdev_nvme.c:2182:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:30:39.463 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.463 17:51:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 263801 00:30:40.849 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (263801) - No such process 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # true 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:30:40.849 { 00:30:40.849 "params": { 00:30:40.849 "name": "Nvme$subsystem", 00:30:40.849 "trtype": "$TEST_TRANSPORT", 00:30:40.849 "traddr": "$NVMF_FIRST_TARGET_IP", 00:30:40.849 "adrfam": "ipv4", 00:30:40.849 "trsvcid": "$NVMF_PORT", 00:30:40.849 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:30:40.849 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:30:40.849 "hdgst": ${hdgst:-false}, 00:30:40.849 "ddgst": ${ddgst:-false} 00:30:40.849 }, 00:30:40.849 "method": "bdev_nvme_attach_controller" 00:30:40.849 } 00:30:40.849 EOF 00:30:40.849 )") 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:30:40.849 17:51:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:30:40.849 "params": { 00:30:40.849 "name": "Nvme0", 00:30:40.849 "trtype": "tcp", 00:30:40.849 "traddr": "10.0.0.2", 00:30:40.849 "adrfam": "ipv4", 00:30:40.849 "trsvcid": "4420", 00:30:40.849 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:30:40.849 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:30:40.849 "hdgst": false, 00:30:40.849 "ddgst": false 00:30:40.849 }, 00:30:40.849 "method": "bdev_nvme_attach_controller" 00:30:40.849 }' 00:30:40.849 [2024-10-17 17:51:48.436121] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:30:40.849 [2024-10-17 17:51:48.436205] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid264152 ] 00:30:40.849 [2024-10-17 17:51:48.518531] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:40.849 [2024-10-17 17:51:48.570139] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:40.849 Running I/O for 1 seconds... 00:30:42.234 1942.00 IOPS, 121.38 MiB/s 00:30:42.234 Latency(us) 00:30:42.234 [2024-10-17T15:51:50.153Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:42.234 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:30:42.234 Verification LBA range: start 0x0 length 0x400 00:30:42.234 Nvme0n1 : 1.01 1986.72 124.17 0.00 0.00 31514.83 781.65 34297.17 00:30:42.234 [2024-10-17T15:51:50.153Z] =================================================================================================================== 00:30:42.234 [2024-10-17T15:51:50.153Z] Total : 1986.72 124.17 0.00 0.00 31514.83 781.65 34297.17 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@514 -- # nvmfcleanup 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:42.234 rmmod nvme_tcp 00:30:42.234 rmmod nvme_fabrics 00:30:42.234 rmmod nvme_keyring 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@515 -- # '[' -n 263514 ']' 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@516 -- # killprocess 263514 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@950 -- # '[' -z 263514 ']' 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@954 -- # kill -0 263514 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@955 -- # uname 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:42.234 17:51:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 263514 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@968 -- # echo 'killing process with pid 263514' 00:30:42.234 killing process with pid 263514 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@969 -- # kill 263514 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@974 -- # wait 263514 00:30:42.234 [2024-10-17 17:51:50.119264] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-save 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-restore 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:42.234 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:42.495 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:42.495 17:51:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:30:44.409 00:30:44.409 real 0m14.778s 00:30:44.409 user 0m19.186s 00:30:44.409 sys 0m7.542s 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:30:44.409 ************************************ 00:30:44.409 END TEST nvmf_host_management 00:30:44.409 ************************************ 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:30:44.409 ************************************ 00:30:44.409 START TEST nvmf_lvol 00:30:44.409 ************************************ 00:30:44.409 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:30:44.671 * Looking for test storage... 00:30:44.671 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1691 -- # lcov --version 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:30:44.671 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:44.671 --rc genhtml_branch_coverage=1 00:30:44.671 --rc genhtml_function_coverage=1 00:30:44.671 --rc genhtml_legend=1 00:30:44.671 --rc geninfo_all_blocks=1 00:30:44.671 --rc geninfo_unexecuted_blocks=1 00:30:44.671 00:30:44.671 ' 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:30:44.671 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:44.671 --rc genhtml_branch_coverage=1 00:30:44.671 --rc genhtml_function_coverage=1 00:30:44.671 --rc genhtml_legend=1 00:30:44.671 --rc geninfo_all_blocks=1 00:30:44.671 --rc geninfo_unexecuted_blocks=1 00:30:44.671 00:30:44.671 ' 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:30:44.671 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:44.671 --rc genhtml_branch_coverage=1 00:30:44.671 --rc genhtml_function_coverage=1 00:30:44.671 --rc genhtml_legend=1 00:30:44.671 --rc geninfo_all_blocks=1 00:30:44.671 --rc geninfo_unexecuted_blocks=1 00:30:44.671 00:30:44.671 ' 00:30:44.671 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:30:44.672 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:44.672 --rc genhtml_branch_coverage=1 00:30:44.672 --rc genhtml_function_coverage=1 00:30:44.672 --rc genhtml_legend=1 00:30:44.672 --rc geninfo_all_blocks=1 00:30:44.672 --rc geninfo_unexecuted_blocks=1 00:30:44.672 00:30:44.672 ' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@474 -- # prepare_net_devs 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@436 -- # local -g is_hw=no 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@438 -- # remove_spdk_ns 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:30:44.672 17:51:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:30:52.813 Found 0000:31:00.0 (0x8086 - 0x159b) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:30:52.813 Found 0000:31:00.1 (0x8086 - 0x159b) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:30:52.813 Found net devices under 0000:31:00.0: cvl_0_0 00:30:52.813 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:30:52.814 Found net devices under 0000:31:00.1: cvl_0_1 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # is_hw=yes 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:52.814 17:51:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:52.814 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:52.814 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.667 ms 00:30:52.814 00:30:52.814 --- 10.0.0.2 ping statistics --- 00:30:52.814 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:52.814 rtt min/avg/max/mdev = 0.667/0.667/0.667/0.000 ms 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:52.814 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:52.814 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.303 ms 00:30:52.814 00:30:52.814 --- 10.0.0.1 ping statistics --- 00:30:52.814 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:52.814 rtt min/avg/max/mdev = 0.303/0.303/0.303/0.000 ms 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@448 -- # return 0 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@724 -- # xtrace_disable 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@507 -- # nvmfpid=268695 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@508 -- # waitforlisten 268695 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x7 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@831 -- # '[' -z 268695 ']' 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:52.814 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:52.814 17:52:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:30:52.814 [2024-10-17 17:52:00.201944] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:30:52.814 [2024-10-17 17:52:00.203106] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:30:52.814 [2024-10-17 17:52:00.203157] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:52.814 [2024-10-17 17:52:00.295420] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:30:52.814 [2024-10-17 17:52:00.349527] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:52.814 [2024-10-17 17:52:00.349579] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:52.814 [2024-10-17 17:52:00.349589] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:52.814 [2024-10-17 17:52:00.349597] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:52.814 [2024-10-17 17:52:00.349603] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:52.814 [2024-10-17 17:52:00.351714] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:30:52.814 [2024-10-17 17:52:00.351828] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:30:52.814 [2024-10-17 17:52:00.351829] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:52.814 [2024-10-17 17:52:00.429820] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:30:52.814 [2024-10-17 17:52:00.430969] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:30:52.814 [2024-10-17 17:52:00.430984] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:30:52.814 [2024-10-17 17:52:00.431206] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@864 -- # return 0 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@730 -- # xtrace_disable 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:30:53.387 [2024-10-17 17:52:01.217085] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:53.387 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:30:53.647 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:30:53.647 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:30:53.909 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:30:53.909 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:30:54.170 17:52:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:30:54.170 17:52:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=dcb572cf-70c3-4b1e-add2-2d1cc0cf49e5 00:30:54.170 17:52:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u dcb572cf-70c3-4b1e-add2-2d1cc0cf49e5 lvol 20 00:30:54.432 17:52:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=d6cd59b8-1f27-41c6-8653-e661e219f2cf 00:30:54.432 17:52:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:30:54.693 17:52:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 d6cd59b8-1f27-41c6-8653-e661e219f2cf 00:30:54.954 17:52:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:30:54.954 [2024-10-17 17:52:02.788953] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:54.954 17:52:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:30:55.215 17:52:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=269225 00:30:55.215 17:52:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:30:55.215 17:52:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:30:56.157 17:52:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot d6cd59b8-1f27-41c6-8653-e661e219f2cf MY_SNAPSHOT 00:30:56.419 17:52:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=f1ef2ac0-4e87-407d-b03d-f67a7e054b78 00:30:56.419 17:52:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize d6cd59b8-1f27-41c6-8653-e661e219f2cf 30 00:30:56.680 17:52:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone f1ef2ac0-4e87-407d-b03d-f67a7e054b78 MY_CLONE 00:30:56.940 17:52:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=a0252122-63ea-42e5-a794-75a930cc44b9 00:30:56.940 17:52:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate a0252122-63ea-42e5-a794-75a930cc44b9 00:30:57.511 17:52:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 269225 00:31:05.642 Initializing NVMe Controllers 00:31:05.642 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:31:05.642 Controller IO queue size 128, less than required. 00:31:05.642 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:31:05.642 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:31:05.642 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:31:05.642 Initialization complete. Launching workers. 00:31:05.642 ======================================================== 00:31:05.642 Latency(us) 00:31:05.642 Device Information : IOPS MiB/s Average min max 00:31:05.642 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 15103.10 59.00 8476.71 1838.23 75101.90 00:31:05.642 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 15079.00 58.90 8490.33 4044.61 70694.42 00:31:05.642 ======================================================== 00:31:05.642 Total : 30182.10 117.90 8483.52 1838.23 75101.90 00:31:05.642 00:31:05.642 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:31:05.642 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete d6cd59b8-1f27-41c6-8653-e661e219f2cf 00:31:05.902 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u dcb572cf-70c3-4b1e-add2-2d1cc0cf49e5 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@514 -- # nvmfcleanup 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:06.162 rmmod nvme_tcp 00:31:06.162 rmmod nvme_fabrics 00:31:06.162 rmmod nvme_keyring 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@515 -- # '[' -n 268695 ']' 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@516 -- # killprocess 268695 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@950 -- # '[' -z 268695 ']' 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@954 -- # kill -0 268695 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@955 -- # uname 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 268695 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@968 -- # echo 'killing process with pid 268695' 00:31:06.162 killing process with pid 268695 00:31:06.162 17:52:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@969 -- # kill 268695 00:31:06.162 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@974 -- # wait 268695 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-save 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-restore 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:06.422 17:52:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:08.333 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:08.333 00:31:08.333 real 0m23.898s 00:31:08.333 user 0m55.682s 00:31:08.333 sys 0m10.890s 00:31:08.333 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:08.333 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:31:08.333 ************************************ 00:31:08.333 END TEST nvmf_lvol 00:31:08.333 ************************************ 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:31:08.594 ************************************ 00:31:08.594 START TEST nvmf_lvs_grow 00:31:08.594 ************************************ 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:31:08.594 * Looking for test storage... 00:31:08.594 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1691 -- # lcov --version 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:31:08.594 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:31:08.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:08.595 --rc genhtml_branch_coverage=1 00:31:08.595 --rc genhtml_function_coverage=1 00:31:08.595 --rc genhtml_legend=1 00:31:08.595 --rc geninfo_all_blocks=1 00:31:08.595 --rc geninfo_unexecuted_blocks=1 00:31:08.595 00:31:08.595 ' 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:31:08.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:08.595 --rc genhtml_branch_coverage=1 00:31:08.595 --rc genhtml_function_coverage=1 00:31:08.595 --rc genhtml_legend=1 00:31:08.595 --rc geninfo_all_blocks=1 00:31:08.595 --rc geninfo_unexecuted_blocks=1 00:31:08.595 00:31:08.595 ' 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:31:08.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:08.595 --rc genhtml_branch_coverage=1 00:31:08.595 --rc genhtml_function_coverage=1 00:31:08.595 --rc genhtml_legend=1 00:31:08.595 --rc geninfo_all_blocks=1 00:31:08.595 --rc geninfo_unexecuted_blocks=1 00:31:08.595 00:31:08.595 ' 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:31:08.595 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:08.595 --rc genhtml_branch_coverage=1 00:31:08.595 --rc genhtml_function_coverage=1 00:31:08.595 --rc genhtml_legend=1 00:31:08.595 --rc geninfo_all_blocks=1 00:31:08.595 --rc geninfo_unexecuted_blocks=1 00:31:08.595 00:31:08.595 ' 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:31:08.595 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:31:08.856 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@474 -- # prepare_net_devs 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@436 -- # local -g is_hw=no 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@438 -- # remove_spdk_ns 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:31:08.857 17:52:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:31:16.995 Found 0000:31:00.0 (0x8086 - 0x159b) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:31:16.995 Found 0000:31:00.1 (0x8086 - 0x159b) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:31:16.995 Found net devices under 0000:31:00.0: cvl_0_0 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:16.995 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:31:16.996 Found net devices under 0000:31:00.1: cvl_0_1 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # is_hw=yes 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:16.996 17:52:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:16.996 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:16.996 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.706 ms 00:31:16.996 00:31:16.996 --- 10.0.0.2 ping statistics --- 00:31:16.996 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:16.996 rtt min/avg/max/mdev = 0.706/0.706/0.706/0.000 ms 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:16.996 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:16.996 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.294 ms 00:31:16.996 00:31:16.996 --- 10.0.0.1 ping statistics --- 00:31:16.996 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:16.996 rtt min/avg/max/mdev = 0.294/0.294/0.294/0.000 ms 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@448 -- # return 0 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@724 -- # xtrace_disable 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@507 -- # nvmfpid=275588 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@508 -- # waitforlisten 275588 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@831 -- # '[' -z 275588 ']' 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:16.996 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:16.996 17:52:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:31:16.996 [2024-10-17 17:52:24.184146] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:31:16.996 [2024-10-17 17:52:24.185328] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:31:16.996 [2024-10-17 17:52:24.185381] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:16.996 [2024-10-17 17:52:24.276685] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:16.996 [2024-10-17 17:52:24.327809] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:16.996 [2024-10-17 17:52:24.327857] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:16.996 [2024-10-17 17:52:24.327866] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:16.996 [2024-10-17 17:52:24.327874] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:16.996 [2024-10-17 17:52:24.327881] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:16.996 [2024-10-17 17:52:24.328659] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:16.996 [2024-10-17 17:52:24.405473] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:31:16.996 [2024-10-17 17:52:24.405768] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:31:17.257 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:17.257 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # return 0 00:31:17.257 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:31:17.257 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@730 -- # xtrace_disable 00:31:17.257 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:31:17.257 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:17.257 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:31:17.518 [2024-10-17 17:52:25.225551] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:31:17.518 ************************************ 00:31:17.518 START TEST lvs_grow_clean 00:31:17.518 ************************************ 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1125 -- # lvs_grow 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:17.518 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:31:17.797 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:31:17.797 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:31:18.092 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:18.092 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:18.092 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:31:18.092 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:31:18.092 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:31:18.092 17:52:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 lvol 150 00:31:18.381 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=2298571c-963b-4c47-aaec-10a992985a84 00:31:18.381 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:18.381 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:31:18.381 [2024-10-17 17:52:26.253221] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:31:18.381 [2024-10-17 17:52:26.253387] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:31:18.381 true 00:31:18.381 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:18.381 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:31:18.642 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:31:18.642 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:31:18.903 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 2298571c-963b-4c47-aaec-10a992985a84 00:31:19.164 17:52:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:31:19.164 [2024-10-17 17:52:26.989944] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:19.164 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=276128 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 276128 /var/tmp/bdevperf.sock 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@831 -- # '[' -z 276128 ']' 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:31:19.426 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:19.426 17:52:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:31:19.426 [2024-10-17 17:52:27.227913] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:31:19.426 [2024-10-17 17:52:27.227980] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid276128 ] 00:31:19.426 [2024-10-17 17:52:27.311275] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:19.687 [2024-10-17 17:52:27.364244] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:20.259 17:52:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:20.259 17:52:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # return 0 00:31:20.259 17:52:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:31:20.520 Nvme0n1 00:31:20.520 17:52:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:31:20.781 [ 00:31:20.781 { 00:31:20.781 "name": "Nvme0n1", 00:31:20.781 "aliases": [ 00:31:20.781 "2298571c-963b-4c47-aaec-10a992985a84" 00:31:20.781 ], 00:31:20.781 "product_name": "NVMe disk", 00:31:20.781 "block_size": 4096, 00:31:20.781 "num_blocks": 38912, 00:31:20.781 "uuid": "2298571c-963b-4c47-aaec-10a992985a84", 00:31:20.781 "numa_id": 0, 00:31:20.781 "assigned_rate_limits": { 00:31:20.781 "rw_ios_per_sec": 0, 00:31:20.781 "rw_mbytes_per_sec": 0, 00:31:20.781 "r_mbytes_per_sec": 0, 00:31:20.781 "w_mbytes_per_sec": 0 00:31:20.781 }, 00:31:20.781 "claimed": false, 00:31:20.781 "zoned": false, 00:31:20.781 "supported_io_types": { 00:31:20.781 "read": true, 00:31:20.781 "write": true, 00:31:20.781 "unmap": true, 00:31:20.781 "flush": true, 00:31:20.781 "reset": true, 00:31:20.781 "nvme_admin": true, 00:31:20.781 "nvme_io": true, 00:31:20.781 "nvme_io_md": false, 00:31:20.781 "write_zeroes": true, 00:31:20.781 "zcopy": false, 00:31:20.781 "get_zone_info": false, 00:31:20.781 "zone_management": false, 00:31:20.781 "zone_append": false, 00:31:20.781 "compare": true, 00:31:20.781 "compare_and_write": true, 00:31:20.781 "abort": true, 00:31:20.781 "seek_hole": false, 00:31:20.781 "seek_data": false, 00:31:20.781 "copy": true, 00:31:20.781 "nvme_iov_md": false 00:31:20.781 }, 00:31:20.781 "memory_domains": [ 00:31:20.781 { 00:31:20.781 "dma_device_id": "system", 00:31:20.781 "dma_device_type": 1 00:31:20.781 } 00:31:20.781 ], 00:31:20.781 "driver_specific": { 00:31:20.781 "nvme": [ 00:31:20.781 { 00:31:20.781 "trid": { 00:31:20.781 "trtype": "TCP", 00:31:20.781 "adrfam": "IPv4", 00:31:20.781 "traddr": "10.0.0.2", 00:31:20.781 "trsvcid": "4420", 00:31:20.781 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:31:20.781 }, 00:31:20.781 "ctrlr_data": { 00:31:20.781 "cntlid": 1, 00:31:20.781 "vendor_id": "0x8086", 00:31:20.781 "model_number": "SPDK bdev Controller", 00:31:20.781 "serial_number": "SPDK0", 00:31:20.781 "firmware_revision": "25.01", 00:31:20.781 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:31:20.781 "oacs": { 00:31:20.781 "security": 0, 00:31:20.781 "format": 0, 00:31:20.781 "firmware": 0, 00:31:20.781 "ns_manage": 0 00:31:20.781 }, 00:31:20.781 "multi_ctrlr": true, 00:31:20.781 "ana_reporting": false 00:31:20.781 }, 00:31:20.781 "vs": { 00:31:20.781 "nvme_version": "1.3" 00:31:20.781 }, 00:31:20.781 "ns_data": { 00:31:20.781 "id": 1, 00:31:20.781 "can_share": true 00:31:20.781 } 00:31:20.781 } 00:31:20.781 ], 00:31:20.781 "mp_policy": "active_passive" 00:31:20.781 } 00:31:20.781 } 00:31:20.781 ] 00:31:20.781 17:52:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=276326 00:31:20.781 17:52:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:31:20.781 17:52:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:31:20.781 Running I/O for 10 seconds... 00:31:22.169 Latency(us) 00:31:22.169 [2024-10-17T15:52:30.088Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:22.169 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:22.169 Nvme0n1 : 1.00 16566.00 64.71 0.00 0.00 0.00 0.00 0.00 00:31:22.169 [2024-10-17T15:52:30.088Z] =================================================================================================================== 00:31:22.169 [2024-10-17T15:52:30.088Z] Total : 16566.00 64.71 0.00 0.00 0.00 0.00 0.00 00:31:22.169 00:31:22.741 17:52:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:23.002 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:23.002 Nvme0n1 : 2.00 16787.00 65.57 0.00 0.00 0.00 0.00 0.00 00:31:23.002 [2024-10-17T15:52:30.921Z] =================================================================================================================== 00:31:23.002 [2024-10-17T15:52:30.921Z] Total : 16787.00 65.57 0.00 0.00 0.00 0.00 0.00 00:31:23.002 00:31:23.002 true 00:31:23.002 17:52:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:23.003 17:52:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:31:23.263 17:52:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:31:23.263 17:52:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:31:23.263 17:52:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 276326 00:31:23.835 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:23.835 Nvme0n1 : 3.00 17056.00 66.62 0.00 0.00 0.00 0.00 0.00 00:31:23.835 [2024-10-17T15:52:31.754Z] =================================================================================================================== 00:31:23.835 [2024-10-17T15:52:31.754Z] Total : 17056.00 66.62 0.00 0.00 0.00 0.00 0.00 00:31:23.835 00:31:24.776 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:24.776 Nvme0n1 : 4.00 17272.00 67.47 0.00 0.00 0.00 0.00 0.00 00:31:24.776 [2024-10-17T15:52:32.695Z] =================================================================================================================== 00:31:24.776 [2024-10-17T15:52:32.695Z] Total : 17272.00 67.47 0.00 0.00 0.00 0.00 0.00 00:31:24.776 00:31:26.161 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:26.161 Nvme0n1 : 5.00 18873.40 73.72 0.00 0.00 0.00 0.00 0.00 00:31:26.161 [2024-10-17T15:52:34.080Z] =================================================================================================================== 00:31:26.161 [2024-10-17T15:52:34.080Z] Total : 18873.40 73.72 0.00 0.00 0.00 0.00 0.00 00:31:26.161 00:31:27.103 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:27.103 Nvme0n1 : 6.00 19941.33 77.90 0.00 0.00 0.00 0.00 0.00 00:31:27.103 [2024-10-17T15:52:35.022Z] =================================================================================================================== 00:31:27.103 [2024-10-17T15:52:35.022Z] Total : 19941.33 77.90 0.00 0.00 0.00 0.00 0.00 00:31:27.103 00:31:28.045 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:28.045 Nvme0n1 : 7.00 20713.00 80.91 0.00 0.00 0.00 0.00 0.00 00:31:28.045 [2024-10-17T15:52:35.964Z] =================================================================================================================== 00:31:28.045 [2024-10-17T15:52:35.964Z] Total : 20713.00 80.91 0.00 0.00 0.00 0.00 0.00 00:31:28.045 00:31:28.988 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:28.988 Nvme0n1 : 8.00 21291.88 83.17 0.00 0.00 0.00 0.00 0.00 00:31:28.988 [2024-10-17T15:52:36.907Z] =================================================================================================================== 00:31:28.988 [2024-10-17T15:52:36.907Z] Total : 21291.88 83.17 0.00 0.00 0.00 0.00 0.00 00:31:28.988 00:31:29.926 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:29.926 Nvme0n1 : 9.00 21742.11 84.93 0.00 0.00 0.00 0.00 0.00 00:31:29.926 [2024-10-17T15:52:37.845Z] =================================================================================================================== 00:31:29.926 [2024-10-17T15:52:37.845Z] Total : 21742.11 84.93 0.00 0.00 0.00 0.00 0.00 00:31:29.926 00:31:30.867 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:30.867 Nvme0n1 : 10.00 22102.30 86.34 0.00 0.00 0.00 0.00 0.00 00:31:30.867 [2024-10-17T15:52:38.786Z] =================================================================================================================== 00:31:30.867 [2024-10-17T15:52:38.786Z] Total : 22102.30 86.34 0.00 0.00 0.00 0.00 0.00 00:31:30.867 00:31:30.867 00:31:30.867 Latency(us) 00:31:30.867 [2024-10-17T15:52:38.786Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:30.867 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:30.867 Nvme0n1 : 10.00 22107.85 86.36 0.00 0.00 5786.31 4177.92 31894.19 00:31:30.867 [2024-10-17T15:52:38.786Z] =================================================================================================================== 00:31:30.867 [2024-10-17T15:52:38.786Z] Total : 22107.85 86.36 0.00 0.00 5786.31 4177.92 31894.19 00:31:30.867 { 00:31:30.867 "results": [ 00:31:30.867 { 00:31:30.867 "job": "Nvme0n1", 00:31:30.867 "core_mask": "0x2", 00:31:30.867 "workload": "randwrite", 00:31:30.867 "status": "finished", 00:31:30.867 "queue_depth": 128, 00:31:30.867 "io_size": 4096, 00:31:30.867 "runtime": 10.00328, 00:31:30.867 "iops": 22107.848625650786, 00:31:30.867 "mibps": 86.35878369394838, 00:31:30.867 "io_failed": 0, 00:31:30.867 "io_timeout": 0, 00:31:30.867 "avg_latency_us": 5786.311510476251, 00:31:30.867 "min_latency_us": 4177.92, 00:31:30.867 "max_latency_us": 31894.18666666667 00:31:30.867 } 00:31:30.867 ], 00:31:30.867 "core_count": 1 00:31:30.867 } 00:31:30.867 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 276128 00:31:30.867 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@950 -- # '[' -z 276128 ']' 00:31:30.867 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # kill -0 276128 00:31:30.867 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # uname 00:31:30.867 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:30.867 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 276128 00:31:31.128 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:31:31.128 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:31:31.128 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 276128' 00:31:31.128 killing process with pid 276128 00:31:31.128 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@969 -- # kill 276128 00:31:31.128 Received shutdown signal, test time was about 10.000000 seconds 00:31:31.128 00:31:31.128 Latency(us) 00:31:31.128 [2024-10-17T15:52:39.047Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:31.128 [2024-10-17T15:52:39.047Z] =================================================================================================================== 00:31:31.128 [2024-10-17T15:52:39.047Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:31:31.128 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@974 -- # wait 276128 00:31:31.128 17:52:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:31:31.388 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:31:31.388 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:31.388 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:31:31.648 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:31:31.649 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:31:31.649 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:31:31.909 [2024-10-17 17:52:39.573271] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@650 -- # local es=0 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:31.909 request: 00:31:31.909 { 00:31:31.909 "uuid": "25642834-a7f0-4cc0-8496-fdf0d2cd9a13", 00:31:31.909 "method": "bdev_lvol_get_lvstores", 00:31:31.909 "req_id": 1 00:31:31.909 } 00:31:31.909 Got JSON-RPC error response 00:31:31.909 response: 00:31:31.909 { 00:31:31.909 "code": -19, 00:31:31.909 "message": "No such device" 00:31:31.909 } 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # es=1 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:31.909 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:31:32.171 aio_bdev 00:31:32.171 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 2298571c-963b-4c47-aaec-10a992985a84 00:31:32.171 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@899 -- # local bdev_name=2298571c-963b-4c47-aaec-10a992985a84 00:31:32.171 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:31:32.171 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@901 -- # local i 00:31:32.171 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:31:32.171 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:31:32.171 17:52:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:32.432 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 2298571c-963b-4c47-aaec-10a992985a84 -t 2000 00:31:32.432 [ 00:31:32.432 { 00:31:32.432 "name": "2298571c-963b-4c47-aaec-10a992985a84", 00:31:32.432 "aliases": [ 00:31:32.432 "lvs/lvol" 00:31:32.432 ], 00:31:32.432 "product_name": "Logical Volume", 00:31:32.432 "block_size": 4096, 00:31:32.432 "num_blocks": 38912, 00:31:32.432 "uuid": "2298571c-963b-4c47-aaec-10a992985a84", 00:31:32.432 "assigned_rate_limits": { 00:31:32.432 "rw_ios_per_sec": 0, 00:31:32.432 "rw_mbytes_per_sec": 0, 00:31:32.432 "r_mbytes_per_sec": 0, 00:31:32.432 "w_mbytes_per_sec": 0 00:31:32.432 }, 00:31:32.432 "claimed": false, 00:31:32.432 "zoned": false, 00:31:32.432 "supported_io_types": { 00:31:32.432 "read": true, 00:31:32.432 "write": true, 00:31:32.432 "unmap": true, 00:31:32.432 "flush": false, 00:31:32.432 "reset": true, 00:31:32.432 "nvme_admin": false, 00:31:32.432 "nvme_io": false, 00:31:32.432 "nvme_io_md": false, 00:31:32.432 "write_zeroes": true, 00:31:32.432 "zcopy": false, 00:31:32.432 "get_zone_info": false, 00:31:32.432 "zone_management": false, 00:31:32.432 "zone_append": false, 00:31:32.432 "compare": false, 00:31:32.432 "compare_and_write": false, 00:31:32.432 "abort": false, 00:31:32.432 "seek_hole": true, 00:31:32.432 "seek_data": true, 00:31:32.432 "copy": false, 00:31:32.432 "nvme_iov_md": false 00:31:32.432 }, 00:31:32.432 "driver_specific": { 00:31:32.432 "lvol": { 00:31:32.432 "lvol_store_uuid": "25642834-a7f0-4cc0-8496-fdf0d2cd9a13", 00:31:32.432 "base_bdev": "aio_bdev", 00:31:32.432 "thin_provision": false, 00:31:32.432 "num_allocated_clusters": 38, 00:31:32.432 "snapshot": false, 00:31:32.432 "clone": false, 00:31:32.432 "esnap_clone": false 00:31:32.432 } 00:31:32.432 } 00:31:32.432 } 00:31:32.432 ] 00:31:32.692 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@907 -- # return 0 00:31:32.693 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:32.693 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:31:32.693 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:31:32.693 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:32.693 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:31:32.953 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:31:32.953 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 2298571c-963b-4c47-aaec-10a992985a84 00:31:32.953 17:52:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 25642834-a7f0-4cc0-8496-fdf0d2cd9a13 00:31:33.214 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:31:33.474 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:33.474 00:31:33.474 real 0m15.942s 00:31:33.474 user 0m15.575s 00:31:33.474 sys 0m1.475s 00:31:33.474 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:33.474 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:31:33.474 ************************************ 00:31:33.474 END TEST lvs_grow_clean 00:31:33.474 ************************************ 00:31:33.474 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:31:33.474 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:31:33.474 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:33.474 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:31:33.474 ************************************ 00:31:33.474 START TEST lvs_grow_dirty 00:31:33.474 ************************************ 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1125 -- # lvs_grow dirty 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:33.475 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:31:33.735 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:31:33.735 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:31:33.996 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:33.996 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:33.996 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:31:33.996 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:31:33.996 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:31:33.996 17:52:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u a6ba1768-e07c-405f-a9ae-04586f8471ae lvol 150 00:31:34.257 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=1af57e76-6bd3-4945-af5b-8434d781a02b 00:31:34.257 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:34.257 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:31:34.519 [2024-10-17 17:52:42.213188] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:31:34.519 [2024-10-17 17:52:42.213332] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:31:34.519 true 00:31:34.519 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:34.519 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:31:34.519 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:31:34.519 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:31:34.779 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 1af57e76-6bd3-4945-af5b-8434d781a02b 00:31:35.040 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:31:35.040 [2024-10-17 17:52:42.857755] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:35.040 17:52:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=279074 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 279074 /var/tmp/bdevperf.sock 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 279074 ']' 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:31:35.301 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:35.301 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:31:35.301 [2024-10-17 17:52:43.095070] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:31:35.301 [2024-10-17 17:52:43.095125] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid279074 ] 00:31:35.301 [2024-10-17 17:52:43.174117] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:35.301 [2024-10-17 17:52:43.203968] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:36.245 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:36.245 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:31:36.245 17:52:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:31:36.245 Nvme0n1 00:31:36.245 17:52:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:31:36.506 [ 00:31:36.506 { 00:31:36.506 "name": "Nvme0n1", 00:31:36.506 "aliases": [ 00:31:36.506 "1af57e76-6bd3-4945-af5b-8434d781a02b" 00:31:36.506 ], 00:31:36.506 "product_name": "NVMe disk", 00:31:36.506 "block_size": 4096, 00:31:36.506 "num_blocks": 38912, 00:31:36.506 "uuid": "1af57e76-6bd3-4945-af5b-8434d781a02b", 00:31:36.506 "numa_id": 0, 00:31:36.506 "assigned_rate_limits": { 00:31:36.506 "rw_ios_per_sec": 0, 00:31:36.506 "rw_mbytes_per_sec": 0, 00:31:36.506 "r_mbytes_per_sec": 0, 00:31:36.506 "w_mbytes_per_sec": 0 00:31:36.506 }, 00:31:36.506 "claimed": false, 00:31:36.506 "zoned": false, 00:31:36.506 "supported_io_types": { 00:31:36.506 "read": true, 00:31:36.506 "write": true, 00:31:36.506 "unmap": true, 00:31:36.506 "flush": true, 00:31:36.506 "reset": true, 00:31:36.506 "nvme_admin": true, 00:31:36.506 "nvme_io": true, 00:31:36.506 "nvme_io_md": false, 00:31:36.506 "write_zeroes": true, 00:31:36.506 "zcopy": false, 00:31:36.506 "get_zone_info": false, 00:31:36.506 "zone_management": false, 00:31:36.506 "zone_append": false, 00:31:36.506 "compare": true, 00:31:36.506 "compare_and_write": true, 00:31:36.506 "abort": true, 00:31:36.506 "seek_hole": false, 00:31:36.507 "seek_data": false, 00:31:36.507 "copy": true, 00:31:36.507 "nvme_iov_md": false 00:31:36.507 }, 00:31:36.507 "memory_domains": [ 00:31:36.507 { 00:31:36.507 "dma_device_id": "system", 00:31:36.507 "dma_device_type": 1 00:31:36.507 } 00:31:36.507 ], 00:31:36.507 "driver_specific": { 00:31:36.507 "nvme": [ 00:31:36.507 { 00:31:36.507 "trid": { 00:31:36.507 "trtype": "TCP", 00:31:36.507 "adrfam": "IPv4", 00:31:36.507 "traddr": "10.0.0.2", 00:31:36.507 "trsvcid": "4420", 00:31:36.507 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:31:36.507 }, 00:31:36.507 "ctrlr_data": { 00:31:36.507 "cntlid": 1, 00:31:36.507 "vendor_id": "0x8086", 00:31:36.507 "model_number": "SPDK bdev Controller", 00:31:36.507 "serial_number": "SPDK0", 00:31:36.507 "firmware_revision": "25.01", 00:31:36.507 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:31:36.507 "oacs": { 00:31:36.507 "security": 0, 00:31:36.507 "format": 0, 00:31:36.507 "firmware": 0, 00:31:36.507 "ns_manage": 0 00:31:36.507 }, 00:31:36.507 "multi_ctrlr": true, 00:31:36.507 "ana_reporting": false 00:31:36.507 }, 00:31:36.507 "vs": { 00:31:36.507 "nvme_version": "1.3" 00:31:36.507 }, 00:31:36.507 "ns_data": { 00:31:36.507 "id": 1, 00:31:36.507 "can_share": true 00:31:36.507 } 00:31:36.507 } 00:31:36.507 ], 00:31:36.507 "mp_policy": "active_passive" 00:31:36.507 } 00:31:36.507 } 00:31:36.507 ] 00:31:36.507 17:52:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=279398 00:31:36.507 17:52:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:31:36.507 17:52:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:31:36.507 Running I/O for 10 seconds... 00:31:37.449 Latency(us) 00:31:37.449 [2024-10-17T15:52:45.368Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:37.449 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:37.449 Nvme0n1 : 1.00 17348.00 67.77 0.00 0.00 0.00 0.00 0.00 00:31:37.449 [2024-10-17T15:52:45.368Z] =================================================================================================================== 00:31:37.449 [2024-10-17T15:52:45.368Z] Total : 17348.00 67.77 0.00 0.00 0.00 0.00 0.00 00:31:37.449 00:31:38.391 17:52:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:38.651 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:38.651 Nvme0n1 : 2.00 17602.50 68.76 0.00 0.00 0.00 0.00 0.00 00:31:38.651 [2024-10-17T15:52:46.570Z] =================================================================================================================== 00:31:38.651 [2024-10-17T15:52:46.570Z] Total : 17602.50 68.76 0.00 0.00 0.00 0.00 0.00 00:31:38.651 00:31:38.651 true 00:31:38.651 17:52:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:38.651 17:52:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:31:38.911 17:52:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:31:38.911 17:52:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:31:38.911 17:52:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 279398 00:31:39.481 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:39.481 Nvme0n1 : 3.00 17686.67 69.09 0.00 0.00 0.00 0.00 0.00 00:31:39.481 [2024-10-17T15:52:47.400Z] =================================================================================================================== 00:31:39.481 [2024-10-17T15:52:47.400Z] Total : 17686.67 69.09 0.00 0.00 0.00 0.00 0.00 00:31:39.481 00:31:40.864 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:40.864 Nvme0n1 : 4.00 17761.25 69.38 0.00 0.00 0.00 0.00 0.00 00:31:40.864 [2024-10-17T15:52:48.783Z] =================================================================================================================== 00:31:40.864 [2024-10-17T15:52:48.783Z] Total : 17761.25 69.38 0.00 0.00 0.00 0.00 0.00 00:31:40.864 00:31:41.804 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:41.804 Nvme0n1 : 5.00 18356.20 71.70 0.00 0.00 0.00 0.00 0.00 00:31:41.804 [2024-10-17T15:52:49.723Z] =================================================================================================================== 00:31:41.804 [2024-10-17T15:52:49.723Z] Total : 18356.20 71.70 0.00 0.00 0.00 0.00 0.00 00:31:41.804 00:31:42.745 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:42.745 Nvme0n1 : 6.00 19499.00 76.17 0.00 0.00 0.00 0.00 0.00 00:31:42.745 [2024-10-17T15:52:50.664Z] =================================================================================================================== 00:31:42.745 [2024-10-17T15:52:50.664Z] Total : 19499.00 76.17 0.00 0.00 0.00 0.00 0.00 00:31:42.745 00:31:43.686 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:43.686 Nvme0n1 : 7.00 20334.00 79.43 0.00 0.00 0.00 0.00 0.00 00:31:43.686 [2024-10-17T15:52:51.605Z] =================================================================================================================== 00:31:43.686 [2024-10-17T15:52:51.605Z] Total : 20334.00 79.43 0.00 0.00 0.00 0.00 0.00 00:31:43.686 00:31:44.626 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:44.626 Nvme0n1 : 8.00 20960.25 81.88 0.00 0.00 0.00 0.00 0.00 00:31:44.626 [2024-10-17T15:52:52.545Z] =================================================================================================================== 00:31:44.626 [2024-10-17T15:52:52.545Z] Total : 20960.25 81.88 0.00 0.00 0.00 0.00 0.00 00:31:44.626 00:31:45.568 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:45.568 Nvme0n1 : 9.00 21454.56 83.81 0.00 0.00 0.00 0.00 0.00 00:31:45.568 [2024-10-17T15:52:53.487Z] =================================================================================================================== 00:31:45.568 [2024-10-17T15:52:53.487Z] Total : 21454.56 83.81 0.00 0.00 0.00 0.00 0.00 00:31:45.568 00:31:46.512 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:46.512 Nvme0n1 : 10.00 21845.20 85.33 0.00 0.00 0.00 0.00 0.00 00:31:46.512 [2024-10-17T15:52:54.431Z] =================================================================================================================== 00:31:46.512 [2024-10-17T15:52:54.431Z] Total : 21845.20 85.33 0.00 0.00 0.00 0.00 0.00 00:31:46.512 00:31:46.512 00:31:46.512 Latency(us) 00:31:46.512 [2024-10-17T15:52:54.431Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:46.512 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:31:46.512 Nvme0n1 : 10.01 21850.75 85.35 0.00 0.00 5854.79 3222.19 23374.51 00:31:46.512 [2024-10-17T15:52:54.431Z] =================================================================================================================== 00:31:46.512 [2024-10-17T15:52:54.431Z] Total : 21850.75 85.35 0.00 0.00 5854.79 3222.19 23374.51 00:31:46.512 { 00:31:46.512 "results": [ 00:31:46.512 { 00:31:46.512 "job": "Nvme0n1", 00:31:46.512 "core_mask": "0x2", 00:31:46.512 "workload": "randwrite", 00:31:46.512 "status": "finished", 00:31:46.512 "queue_depth": 128, 00:31:46.512 "io_size": 4096, 00:31:46.512 "runtime": 10.005376, 00:31:46.512 "iops": 21850.753035168294, 00:31:46.512 "mibps": 85.35450404362615, 00:31:46.512 "io_failed": 0, 00:31:46.512 "io_timeout": 0, 00:31:46.512 "avg_latency_us": 5854.788676500858, 00:31:46.512 "min_latency_us": 3222.1866666666665, 00:31:46.512 "max_latency_us": 23374.506666666668 00:31:46.512 } 00:31:46.512 ], 00:31:46.512 "core_count": 1 00:31:46.512 } 00:31:46.512 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 279074 00:31:46.512 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@950 -- # '[' -z 279074 ']' 00:31:46.512 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # kill -0 279074 00:31:46.512 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # uname 00:31:46.512 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:46.512 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 279074 00:31:46.774 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:31:46.774 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:31:46.774 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@968 -- # echo 'killing process with pid 279074' 00:31:46.774 killing process with pid 279074 00:31:46.774 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@969 -- # kill 279074 00:31:46.774 Received shutdown signal, test time was about 10.000000 seconds 00:31:46.774 00:31:46.774 Latency(us) 00:31:46.774 [2024-10-17T15:52:54.693Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:46.774 [2024-10-17T15:52:54.693Z] =================================================================================================================== 00:31:46.774 [2024-10-17T15:52:54.693Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:31:46.774 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@974 -- # wait 279074 00:31:46.774 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:31:47.035 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:31:47.035 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:47.035 17:52:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 275588 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 275588 00:31:47.296 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 275588 Killed "${NVMF_APP[@]}" "$@" 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@724 -- # xtrace_disable 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # nvmfpid=281411 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # waitforlisten 281411 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 281411 ']' 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:47.296 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:47.296 17:52:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:31:47.296 [2024-10-17 17:52:55.200722] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:31:47.296 [2024-10-17 17:52:55.201768] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:31:47.296 [2024-10-17 17:52:55.201813] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:47.556 [2024-10-17 17:52:55.288189] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:47.556 [2024-10-17 17:52:55.320357] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:47.556 [2024-10-17 17:52:55.320386] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:47.556 [2024-10-17 17:52:55.320392] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:47.556 [2024-10-17 17:52:55.320396] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:47.557 [2024-10-17 17:52:55.320401] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:47.557 [2024-10-17 17:52:55.320882] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:47.557 [2024-10-17 17:52:55.372326] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:31:47.557 [2024-10-17 17:52:55.372531] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:31:48.128 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:48.128 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:31:48.128 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:31:48.128 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@730 -- # xtrace_disable 00:31:48.128 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:31:48.128 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:48.128 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:31:48.388 [2024-10-17 17:52:56.219339] blobstore.c:4875:bs_recover: *NOTICE*: Performing recovery on blobstore 00:31:48.388 [2024-10-17 17:52:56.219644] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:31:48.388 [2024-10-17 17:52:56.219746] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev 1af57e76-6bd3-4945-af5b-8434d781a02b 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=1af57e76-6bd3-4945-af5b-8434d781a02b 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:31:48.388 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:48.647 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 1af57e76-6bd3-4945-af5b-8434d781a02b -t 2000 00:31:48.906 [ 00:31:48.906 { 00:31:48.906 "name": "1af57e76-6bd3-4945-af5b-8434d781a02b", 00:31:48.906 "aliases": [ 00:31:48.906 "lvs/lvol" 00:31:48.906 ], 00:31:48.906 "product_name": "Logical Volume", 00:31:48.906 "block_size": 4096, 00:31:48.906 "num_blocks": 38912, 00:31:48.906 "uuid": "1af57e76-6bd3-4945-af5b-8434d781a02b", 00:31:48.906 "assigned_rate_limits": { 00:31:48.906 "rw_ios_per_sec": 0, 00:31:48.906 "rw_mbytes_per_sec": 0, 00:31:48.906 "r_mbytes_per_sec": 0, 00:31:48.906 "w_mbytes_per_sec": 0 00:31:48.906 }, 00:31:48.906 "claimed": false, 00:31:48.906 "zoned": false, 00:31:48.906 "supported_io_types": { 00:31:48.906 "read": true, 00:31:48.906 "write": true, 00:31:48.906 "unmap": true, 00:31:48.906 "flush": false, 00:31:48.906 "reset": true, 00:31:48.906 "nvme_admin": false, 00:31:48.906 "nvme_io": false, 00:31:48.906 "nvme_io_md": false, 00:31:48.906 "write_zeroes": true, 00:31:48.906 "zcopy": false, 00:31:48.906 "get_zone_info": false, 00:31:48.906 "zone_management": false, 00:31:48.906 "zone_append": false, 00:31:48.906 "compare": false, 00:31:48.906 "compare_and_write": false, 00:31:48.906 "abort": false, 00:31:48.906 "seek_hole": true, 00:31:48.906 "seek_data": true, 00:31:48.906 "copy": false, 00:31:48.906 "nvme_iov_md": false 00:31:48.906 }, 00:31:48.906 "driver_specific": { 00:31:48.906 "lvol": { 00:31:48.906 "lvol_store_uuid": "a6ba1768-e07c-405f-a9ae-04586f8471ae", 00:31:48.906 "base_bdev": "aio_bdev", 00:31:48.906 "thin_provision": false, 00:31:48.906 "num_allocated_clusters": 38, 00:31:48.906 "snapshot": false, 00:31:48.906 "clone": false, 00:31:48.906 "esnap_clone": false 00:31:48.906 } 00:31:48.906 } 00:31:48.906 } 00:31:48.906 ] 00:31:48.906 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:31:48.906 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:48.906 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:31:48.906 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:31:48.907 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:48.907 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:31:49.166 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:31:49.166 17:52:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:31:49.166 [2024-10-17 17:52:57.081370] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@650 -- # local es=0 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:49.426 request: 00:31:49.426 { 00:31:49.426 "uuid": "a6ba1768-e07c-405f-a9ae-04586f8471ae", 00:31:49.426 "method": "bdev_lvol_get_lvstores", 00:31:49.426 "req_id": 1 00:31:49.426 } 00:31:49.426 Got JSON-RPC error response 00:31:49.426 response: 00:31:49.426 { 00:31:49.426 "code": -19, 00:31:49.426 "message": "No such device" 00:31:49.426 } 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # es=1 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:49.426 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:31:49.686 aio_bdev 00:31:49.686 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 1af57e76-6bd3-4945-af5b-8434d781a02b 00:31:49.686 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=1af57e76-6bd3-4945-af5b-8434d781a02b 00:31:49.686 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:31:49.686 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:31:49.686 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:31:49.686 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:31:49.686 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:31:49.948 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 1af57e76-6bd3-4945-af5b-8434d781a02b -t 2000 00:31:49.948 [ 00:31:49.948 { 00:31:49.948 "name": "1af57e76-6bd3-4945-af5b-8434d781a02b", 00:31:49.948 "aliases": [ 00:31:49.948 "lvs/lvol" 00:31:49.948 ], 00:31:49.948 "product_name": "Logical Volume", 00:31:49.948 "block_size": 4096, 00:31:49.948 "num_blocks": 38912, 00:31:49.948 "uuid": "1af57e76-6bd3-4945-af5b-8434d781a02b", 00:31:49.948 "assigned_rate_limits": { 00:31:49.948 "rw_ios_per_sec": 0, 00:31:49.948 "rw_mbytes_per_sec": 0, 00:31:49.948 "r_mbytes_per_sec": 0, 00:31:49.948 "w_mbytes_per_sec": 0 00:31:49.948 }, 00:31:49.948 "claimed": false, 00:31:49.948 "zoned": false, 00:31:49.948 "supported_io_types": { 00:31:49.948 "read": true, 00:31:49.948 "write": true, 00:31:49.948 "unmap": true, 00:31:49.948 "flush": false, 00:31:49.948 "reset": true, 00:31:49.948 "nvme_admin": false, 00:31:49.948 "nvme_io": false, 00:31:49.948 "nvme_io_md": false, 00:31:49.948 "write_zeroes": true, 00:31:49.948 "zcopy": false, 00:31:49.948 "get_zone_info": false, 00:31:49.948 "zone_management": false, 00:31:49.948 "zone_append": false, 00:31:49.948 "compare": false, 00:31:49.948 "compare_and_write": false, 00:31:49.948 "abort": false, 00:31:49.948 "seek_hole": true, 00:31:49.948 "seek_data": true, 00:31:49.948 "copy": false, 00:31:49.948 "nvme_iov_md": false 00:31:49.948 }, 00:31:49.948 "driver_specific": { 00:31:49.948 "lvol": { 00:31:49.948 "lvol_store_uuid": "a6ba1768-e07c-405f-a9ae-04586f8471ae", 00:31:49.948 "base_bdev": "aio_bdev", 00:31:49.948 "thin_provision": false, 00:31:49.948 "num_allocated_clusters": 38, 00:31:49.948 "snapshot": false, 00:31:49.948 "clone": false, 00:31:49.948 "esnap_clone": false 00:31:49.948 } 00:31:49.948 } 00:31:49.948 } 00:31:49.948 ] 00:31:49.948 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:31:49.948 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:49.948 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:31:50.209 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:31:50.209 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:50.209 17:52:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:31:50.469 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:31:50.469 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 1af57e76-6bd3-4945-af5b-8434d781a02b 00:31:50.469 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u a6ba1768-e07c-405f-a9ae-04586f8471ae 00:31:50.729 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:31:50.990 00:31:50.990 real 0m17.404s 00:31:50.990 user 0m35.227s 00:31:50.990 sys 0m3.187s 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:31:50.990 ************************************ 00:31:50.990 END TEST lvs_grow_dirty 00:31:50.990 ************************************ 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@808 -- # type=--id 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@809 -- # id=0 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # for n in $shm_files 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:31:50.990 nvmf_trace.0 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@823 -- # return 0 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@514 -- # nvmfcleanup 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:50.990 rmmod nvme_tcp 00:31:50.990 rmmod nvme_fabrics 00:31:50.990 rmmod nvme_keyring 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@515 -- # '[' -n 281411 ']' 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@516 -- # killprocess 281411 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@950 -- # '[' -z 281411 ']' 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # kill -0 281411 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # uname 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:50.990 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 281411 00:31:51.250 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:31:51.250 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:31:51.250 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@968 -- # echo 'killing process with pid 281411' 00:31:51.250 killing process with pid 281411 00:31:51.250 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@969 -- # kill 281411 00:31:51.250 17:52:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@974 -- # wait 281411 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-save 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-restore 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:51.250 17:52:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:53.796 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:53.797 00:31:53.797 real 0m44.853s 00:31:53.797 user 0m53.871s 00:31:53.797 sys 0m10.819s 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:31:53.797 ************************************ 00:31:53.797 END TEST nvmf_lvs_grow 00:31:53.797 ************************************ 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:31:53.797 ************************************ 00:31:53.797 START TEST nvmf_bdev_io_wait 00:31:53.797 ************************************ 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:31:53.797 * Looking for test storage... 00:31:53.797 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1691 -- # lcov --version 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:31:53.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:53.797 --rc genhtml_branch_coverage=1 00:31:53.797 --rc genhtml_function_coverage=1 00:31:53.797 --rc genhtml_legend=1 00:31:53.797 --rc geninfo_all_blocks=1 00:31:53.797 --rc geninfo_unexecuted_blocks=1 00:31:53.797 00:31:53.797 ' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:31:53.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:53.797 --rc genhtml_branch_coverage=1 00:31:53.797 --rc genhtml_function_coverage=1 00:31:53.797 --rc genhtml_legend=1 00:31:53.797 --rc geninfo_all_blocks=1 00:31:53.797 --rc geninfo_unexecuted_blocks=1 00:31:53.797 00:31:53.797 ' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:31:53.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:53.797 --rc genhtml_branch_coverage=1 00:31:53.797 --rc genhtml_function_coverage=1 00:31:53.797 --rc genhtml_legend=1 00:31:53.797 --rc geninfo_all_blocks=1 00:31:53.797 --rc geninfo_unexecuted_blocks=1 00:31:53.797 00:31:53.797 ' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:31:53.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:53.797 --rc genhtml_branch_coverage=1 00:31:53.797 --rc genhtml_function_coverage=1 00:31:53.797 --rc genhtml_legend=1 00:31:53.797 --rc geninfo_all_blocks=1 00:31:53.797 --rc geninfo_unexecuted_blocks=1 00:31:53.797 00:31:53.797 ' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:53.797 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # prepare_net_devs 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@436 -- # local -g is_hw=no 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # remove_spdk_ns 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:31:53.798 17:53:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:32:01.939 Found 0000:31:00.0 (0x8086 - 0x159b) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:32:01.939 Found 0000:31:00.1 (0x8086 - 0x159b) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:32:01.939 Found net devices under 0000:31:00.0: cvl_0_0 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:32:01.939 Found net devices under 0000:31:00.1: cvl_0_1 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # is_hw=yes 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:01.939 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:01.940 17:53:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:01.940 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:01.940 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.624 ms 00:32:01.940 00:32:01.940 --- 10.0.0.2 ping statistics --- 00:32:01.940 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:01.940 rtt min/avg/max/mdev = 0.624/0.624/0.624/0.000 ms 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:01.940 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:01.940 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.276 ms 00:32:01.940 00:32:01.940 --- 10.0.0.1 ping statistics --- 00:32:01.940 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:01.940 rtt min/avg/max/mdev = 0.276/0.276/0.276/0.000 ms 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@448 -- # return 0 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@724 -- # xtrace_disable 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # nvmfpid=286496 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # waitforlisten 286496 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF --wait-for-rpc 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@831 -- # '[' -z 286496 ']' 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:01.940 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:01.940 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:01.940 [2024-10-17 17:53:09.151249] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:32:01.940 [2024-10-17 17:53:09.152420] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:01.940 [2024-10-17 17:53:09.152469] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:01.940 [2024-10-17 17:53:09.243417] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:32:01.940 [2024-10-17 17:53:09.298066] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:01.940 [2024-10-17 17:53:09.298117] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:01.940 [2024-10-17 17:53:09.298125] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:01.940 [2024-10-17 17:53:09.298132] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:01.940 [2024-10-17 17:53:09.298138] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:01.940 [2024-10-17 17:53:09.300227] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:01.940 [2024-10-17 17:53:09.300388] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:01.940 [2024-10-17 17:53:09.300545] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:01.940 [2024-10-17 17:53:09.300546] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:01.940 [2024-10-17 17:53:09.300899] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:32:02.201 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:02.201 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # return 0 00:32:02.201 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:32:02.201 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@730 -- # xtrace_disable 00:32:02.201 17:53:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.201 [2024-10-17 17:53:10.074948] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:32:02.201 [2024-10-17 17:53:10.075642] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:32:02.201 [2024-10-17 17:53:10.075764] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:32:02.201 [2024-10-17 17:53:10.075990] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.201 [2024-10-17 17:53:10.085110] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:02.201 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.463 Malloc0 00:32:02.463 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:02.463 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:02.464 [2024-10-17 17:53:10.161512] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=286547 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=286550 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:02.464 { 00:32:02.464 "params": { 00:32:02.464 "name": "Nvme$subsystem", 00:32:02.464 "trtype": "$TEST_TRANSPORT", 00:32:02.464 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:02.464 "adrfam": "ipv4", 00:32:02.464 "trsvcid": "$NVMF_PORT", 00:32:02.464 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:02.464 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:02.464 "hdgst": ${hdgst:-false}, 00:32:02.464 "ddgst": ${ddgst:-false} 00:32:02.464 }, 00:32:02.464 "method": "bdev_nvme_attach_controller" 00:32:02.464 } 00:32:02.464 EOF 00:32:02.464 )") 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=286552 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=286555 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:02.464 { 00:32:02.464 "params": { 00:32:02.464 "name": "Nvme$subsystem", 00:32:02.464 "trtype": "$TEST_TRANSPORT", 00:32:02.464 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:02.464 "adrfam": "ipv4", 00:32:02.464 "trsvcid": "$NVMF_PORT", 00:32:02.464 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:02.464 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:02.464 "hdgst": ${hdgst:-false}, 00:32:02.464 "ddgst": ${ddgst:-false} 00:32:02.464 }, 00:32:02.464 "method": "bdev_nvme_attach_controller" 00:32:02.464 } 00:32:02.464 EOF 00:32:02.464 )") 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:02.464 { 00:32:02.464 "params": { 00:32:02.464 "name": "Nvme$subsystem", 00:32:02.464 "trtype": "$TEST_TRANSPORT", 00:32:02.464 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:02.464 "adrfam": "ipv4", 00:32:02.464 "trsvcid": "$NVMF_PORT", 00:32:02.464 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:02.464 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:02.464 "hdgst": ${hdgst:-false}, 00:32:02.464 "ddgst": ${ddgst:-false} 00:32:02.464 }, 00:32:02.464 "method": "bdev_nvme_attach_controller" 00:32:02.464 } 00:32:02.464 EOF 00:32:02.464 )") 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:02.464 { 00:32:02.464 "params": { 00:32:02.464 "name": "Nvme$subsystem", 00:32:02.464 "trtype": "$TEST_TRANSPORT", 00:32:02.464 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:02.464 "adrfam": "ipv4", 00:32:02.464 "trsvcid": "$NVMF_PORT", 00:32:02.464 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:02.464 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:02.464 "hdgst": ${hdgst:-false}, 00:32:02.464 "ddgst": ${ddgst:-false} 00:32:02.464 }, 00:32:02.464 "method": "bdev_nvme_attach_controller" 00:32:02.464 } 00:32:02.464 EOF 00:32:02.464 )") 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 286547 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:02.464 "params": { 00:32:02.464 "name": "Nvme1", 00:32:02.464 "trtype": "tcp", 00:32:02.464 "traddr": "10.0.0.2", 00:32:02.464 "adrfam": "ipv4", 00:32:02.464 "trsvcid": "4420", 00:32:02.464 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:02.464 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:02.464 "hdgst": false, 00:32:02.464 "ddgst": false 00:32:02.464 }, 00:32:02.464 "method": "bdev_nvme_attach_controller" 00:32:02.464 }' 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:02.464 "params": { 00:32:02.464 "name": "Nvme1", 00:32:02.464 "trtype": "tcp", 00:32:02.464 "traddr": "10.0.0.2", 00:32:02.464 "adrfam": "ipv4", 00:32:02.464 "trsvcid": "4420", 00:32:02.464 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:02.464 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:02.464 "hdgst": false, 00:32:02.464 "ddgst": false 00:32:02.464 }, 00:32:02.464 "method": "bdev_nvme_attach_controller" 00:32:02.464 }' 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:32:02.464 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:02.464 "params": { 00:32:02.464 "name": "Nvme1", 00:32:02.464 "trtype": "tcp", 00:32:02.464 "traddr": "10.0.0.2", 00:32:02.464 "adrfam": "ipv4", 00:32:02.464 "trsvcid": "4420", 00:32:02.464 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:02.464 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:02.464 "hdgst": false, 00:32:02.464 "ddgst": false 00:32:02.464 }, 00:32:02.464 "method": "bdev_nvme_attach_controller" 00:32:02.464 }' 00:32:02.465 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:32:02.465 17:53:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:02.465 "params": { 00:32:02.465 "name": "Nvme1", 00:32:02.465 "trtype": "tcp", 00:32:02.465 "traddr": "10.0.0.2", 00:32:02.465 "adrfam": "ipv4", 00:32:02.465 "trsvcid": "4420", 00:32:02.465 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:02.465 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:02.465 "hdgst": false, 00:32:02.465 "ddgst": false 00:32:02.465 }, 00:32:02.465 "method": "bdev_nvme_attach_controller" 00:32:02.465 }' 00:32:02.465 [2024-10-17 17:53:10.217394] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:02.465 [2024-10-17 17:53:10.217468] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:32:02.465 [2024-10-17 17:53:10.221411] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:02.465 [2024-10-17 17:53:10.221473] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:32:02.465 [2024-10-17 17:53:10.221829] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:02.465 [2024-10-17 17:53:10.221887] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:32:02.465 [2024-10-17 17:53:10.223020] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:02.465 [2024-10-17 17:53:10.223084] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:32:02.726 [2024-10-17 17:53:10.422432] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:02.726 [2024-10-17 17:53:10.465258] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:32:02.726 [2024-10-17 17:53:10.512343] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:02.726 [2024-10-17 17:53:10.553882] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:32:02.726 [2024-10-17 17:53:10.587142] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:02.726 [2024-10-17 17:53:10.627148] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:32:02.987 [2024-10-17 17:53:10.655895] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:02.987 [2024-10-17 17:53:10.694331] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:32:02.987 Running I/O for 1 seconds... 00:32:02.987 Running I/O for 1 seconds... 00:32:03.248 Running I/O for 1 seconds... 00:32:03.248 Running I/O for 1 seconds... 00:32:04.191 10684.00 IOPS, 41.73 MiB/s 00:32:04.191 Latency(us) 00:32:04.191 [2024-10-17T15:53:12.110Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:04.191 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:32:04.191 Nvme1n1 : 1.01 10743.02 41.96 0.00 0.00 11870.00 2334.72 13926.40 00:32:04.191 [2024-10-17T15:53:12.110Z] =================================================================================================================== 00:32:04.191 [2024-10-17T15:53:12.110Z] Total : 10743.02 41.96 0.00 0.00 11870.00 2334.72 13926.40 00:32:04.191 187824.00 IOPS, 733.69 MiB/s 00:32:04.191 Latency(us) 00:32:04.191 [2024-10-17T15:53:12.110Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:04.191 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:32:04.191 Nvme1n1 : 1.00 187454.47 732.24 0.00 0.00 678.71 302.08 1966.08 00:32:04.191 [2024-10-17T15:53:12.110Z] =================================================================================================================== 00:32:04.191 [2024-10-17T15:53:12.110Z] Total : 187454.47 732.24 0.00 0.00 678.71 302.08 1966.08 00:32:04.191 10355.00 IOPS, 40.45 MiB/s [2024-10-17T15:53:12.110Z] 9019.00 IOPS, 35.23 MiB/s 00:32:04.191 Latency(us) 00:32:04.191 [2024-10-17T15:53:12.110Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:04.191 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:32:04.191 Nvme1n1 : 1.01 10423.43 40.72 0.00 0.00 12236.12 2648.75 17803.95 00:32:04.191 [2024-10-17T15:53:12.110Z] =================================================================================================================== 00:32:04.191 [2024-10-17T15:53:12.110Z] Total : 10423.43 40.72 0.00 0.00 12236.12 2648.75 17803.95 00:32:04.191 00:32:04.191 Latency(us) 00:32:04.191 [2024-10-17T15:53:12.110Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:04.191 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:32:04.191 Nvme1n1 : 1.01 9085.78 35.49 0.00 0.00 14035.31 5925.55 20971.52 00:32:04.191 [2024-10-17T15:53:12.110Z] =================================================================================================================== 00:32:04.191 [2024-10-17T15:53:12.110Z] Total : 9085.78 35.49 0.00 0.00 14035.31 5925.55 20971.52 00:32:04.191 17:53:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 286550 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 286552 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 286555 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@514 -- # nvmfcleanup 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:04.191 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:04.191 rmmod nvme_tcp 00:32:04.191 rmmod nvme_fabrics 00:32:04.452 rmmod nvme_keyring 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@515 -- # '[' -n 286496 ']' 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # killprocess 286496 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@950 -- # '[' -z 286496 ']' 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # kill -0 286496 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # uname 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 286496 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@968 -- # echo 'killing process with pid 286496' 00:32:04.453 killing process with pid 286496 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@969 -- # kill 286496 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@974 -- # wait 286496 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-save 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:32:04.453 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-restore 00:32:04.714 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:04.714 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:04.714 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:04.714 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:04.714 17:53:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:06.737 00:32:06.737 real 0m13.221s 00:32:06.737 user 0m15.865s 00:32:06.737 sys 0m7.988s 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:32:06.737 ************************************ 00:32:06.737 END TEST nvmf_bdev_io_wait 00:32:06.737 ************************************ 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:32:06.737 ************************************ 00:32:06.737 START TEST nvmf_queue_depth 00:32:06.737 ************************************ 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:32:06.737 * Looking for test storage... 00:32:06.737 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1691 -- # lcov --version 00:32:06.737 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:32:06.999 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:06.999 --rc genhtml_branch_coverage=1 00:32:06.999 --rc genhtml_function_coverage=1 00:32:06.999 --rc genhtml_legend=1 00:32:06.999 --rc geninfo_all_blocks=1 00:32:06.999 --rc geninfo_unexecuted_blocks=1 00:32:06.999 00:32:06.999 ' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:32:06.999 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:06.999 --rc genhtml_branch_coverage=1 00:32:06.999 --rc genhtml_function_coverage=1 00:32:06.999 --rc genhtml_legend=1 00:32:06.999 --rc geninfo_all_blocks=1 00:32:06.999 --rc geninfo_unexecuted_blocks=1 00:32:06.999 00:32:06.999 ' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:32:06.999 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:06.999 --rc genhtml_branch_coverage=1 00:32:06.999 --rc genhtml_function_coverage=1 00:32:06.999 --rc genhtml_legend=1 00:32:06.999 --rc geninfo_all_blocks=1 00:32:06.999 --rc geninfo_unexecuted_blocks=1 00:32:06.999 00:32:06.999 ' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:32:06.999 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:06.999 --rc genhtml_branch_coverage=1 00:32:06.999 --rc genhtml_function_coverage=1 00:32:06.999 --rc genhtml_legend=1 00:32:06.999 --rc geninfo_all_blocks=1 00:32:06.999 --rc geninfo_unexecuted_blocks=1 00:32:06.999 00:32:06.999 ' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:06.999 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@474 -- # prepare_net_devs 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@436 -- # local -g is_hw=no 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@438 -- # remove_spdk_ns 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:32:07.000 17:53:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:32:15.146 Found 0000:31:00.0 (0x8086 - 0x159b) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:32:15.146 Found 0000:31:00.1 (0x8086 - 0x159b) 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:15.146 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:32:15.147 Found net devices under 0000:31:00.0: cvl_0_0 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:32:15.147 Found net devices under 0000:31:00.1: cvl_0_1 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # is_hw=yes 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:15.147 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:15.147 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.591 ms 00:32:15.147 00:32:15.147 --- 10.0.0.2 ping statistics --- 00:32:15.147 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:15.147 rtt min/avg/max/mdev = 0.591/0.591/0.591/0.000 ms 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:15.147 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:15.147 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.296 ms 00:32:15.147 00:32:15.147 --- 10.0.0.1 ping statistics --- 00:32:15.147 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:15.147 rtt min/avg/max/mdev = 0.296/0.296/0.296/0.000 ms 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@448 -- # return 0 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@724 -- # xtrace_disable 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@507 -- # nvmfpid=291256 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@508 -- # waitforlisten 291256 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 291256 ']' 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:15.147 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:15.147 17:53:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.147 [2024-10-17 17:53:22.526847] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:32:15.147 [2024-10-17 17:53:22.527981] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:15.147 [2024-10-17 17:53:22.528029] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:15.147 [2024-10-17 17:53:22.621978] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:15.147 [2024-10-17 17:53:22.672441] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:15.147 [2024-10-17 17:53:22.672488] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:15.147 [2024-10-17 17:53:22.672497] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:15.147 [2024-10-17 17:53:22.672504] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:15.147 [2024-10-17 17:53:22.672510] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:15.147 [2024-10-17 17:53:22.673356] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:15.147 [2024-10-17 17:53:22.750647] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:32:15.147 [2024-10-17 17:53:22.750954] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@730 -- # xtrace_disable 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.721 [2024-10-17 17:53:23.390230] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.721 Malloc0 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.721 [2024-10-17 17:53:23.478302] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=291457 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 291457 /var/tmp/bdevperf.sock 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 291457 ']' 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:32:15.721 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:15.721 17:53:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:15.721 [2024-10-17 17:53:23.535397] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:15.721 [2024-10-17 17:53:23.535462] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid291457 ] 00:32:15.722 [2024-10-17 17:53:23.618523] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:15.983 [2024-10-17 17:53:23.671951] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:16.555 17:53:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:16.555 17:53:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:32:16.555 17:53:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:32:16.555 17:53:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:16.555 17:53:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:16.815 NVMe0n1 00:32:16.815 17:53:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:16.815 17:53:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:32:16.815 Running I/O for 10 seconds... 00:32:18.697 8813.00 IOPS, 34.43 MiB/s [2024-10-17T15:53:27.998Z] 8987.50 IOPS, 35.11 MiB/s [2024-10-17T15:53:28.991Z] 9591.00 IOPS, 37.46 MiB/s [2024-10-17T15:53:29.928Z] 10752.00 IOPS, 42.00 MiB/s [2024-10-17T15:53:30.868Z] 11368.00 IOPS, 44.41 MiB/s [2024-10-17T15:53:31.807Z] 11791.67 IOPS, 46.06 MiB/s [2024-10-17T15:53:32.749Z] 12144.29 IOPS, 47.44 MiB/s [2024-10-17T15:53:33.689Z] 12344.62 IOPS, 48.22 MiB/s [2024-10-17T15:53:35.071Z] 12527.89 IOPS, 48.94 MiB/s [2024-10-17T15:53:35.071Z] 12707.20 IOPS, 49.64 MiB/s 00:32:27.152 Latency(us) 00:32:27.152 [2024-10-17T15:53:35.071Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:27.152 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:32:27.152 Verification LBA range: start 0x0 length 0x4000 00:32:27.153 NVMe0n1 : 10.06 12730.08 49.73 0.00 0.00 80173.67 25340.59 66409.81 00:32:27.153 [2024-10-17T15:53:35.072Z] =================================================================================================================== 00:32:27.153 [2024-10-17T15:53:35.072Z] Total : 12730.08 49.73 0.00 0.00 80173.67 25340.59 66409.81 00:32:27.153 { 00:32:27.153 "results": [ 00:32:27.153 { 00:32:27.153 "job": "NVMe0n1", 00:32:27.153 "core_mask": "0x1", 00:32:27.153 "workload": "verify", 00:32:27.153 "status": "finished", 00:32:27.153 "verify_range": { 00:32:27.153 "start": 0, 00:32:27.153 "length": 16384 00:32:27.153 }, 00:32:27.153 "queue_depth": 1024, 00:32:27.153 "io_size": 4096, 00:32:27.153 "runtime": 10.060189, 00:32:27.153 "iops": 12730.078927940618, 00:32:27.153 "mibps": 49.72687081226804, 00:32:27.153 "io_failed": 0, 00:32:27.153 "io_timeout": 0, 00:32:27.153 "avg_latency_us": 80173.66801564806, 00:32:27.153 "min_latency_us": 25340.586666666666, 00:32:27.153 "max_latency_us": 66409.81333333334 00:32:27.153 } 00:32:27.153 ], 00:32:27.153 "core_count": 1 00:32:27.153 } 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 291457 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 291457 ']' 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 291457 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 291457 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 291457' 00:32:27.153 killing process with pid 291457 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 291457 00:32:27.153 Received shutdown signal, test time was about 10.000000 seconds 00:32:27.153 00:32:27.153 Latency(us) 00:32:27.153 [2024-10-17T15:53:35.072Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:27.153 [2024-10-17T15:53:35.072Z] =================================================================================================================== 00:32:27.153 [2024-10-17T15:53:35.072Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 291457 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@514 -- # nvmfcleanup 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:27.153 rmmod nvme_tcp 00:32:27.153 rmmod nvme_fabrics 00:32:27.153 rmmod nvme_keyring 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@515 -- # '[' -n 291256 ']' 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@516 -- # killprocess 291256 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 291256 ']' 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 291256 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:27.153 17:53:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 291256 00:32:27.153 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:27.153 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:27.153 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 291256' 00:32:27.153 killing process with pid 291256 00:32:27.153 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 291256 00:32:27.153 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 291256 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-save 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-restore 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:27.413 17:53:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:29.323 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:29.323 00:32:29.323 real 0m22.676s 00:32:29.323 user 0m24.658s 00:32:29.323 sys 0m7.659s 00:32:29.323 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:29.323 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:32:29.323 ************************************ 00:32:29.323 END TEST nvmf_queue_depth 00:32:29.323 ************************************ 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:32:29.584 ************************************ 00:32:29.584 START TEST nvmf_target_multipath 00:32:29.584 ************************************ 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:32:29.584 * Looking for test storage... 00:32:29.584 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1691 -- # lcov --version 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:32:29.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:29.584 --rc genhtml_branch_coverage=1 00:32:29.584 --rc genhtml_function_coverage=1 00:32:29.584 --rc genhtml_legend=1 00:32:29.584 --rc geninfo_all_blocks=1 00:32:29.584 --rc geninfo_unexecuted_blocks=1 00:32:29.584 00:32:29.584 ' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:32:29.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:29.584 --rc genhtml_branch_coverage=1 00:32:29.584 --rc genhtml_function_coverage=1 00:32:29.584 --rc genhtml_legend=1 00:32:29.584 --rc geninfo_all_blocks=1 00:32:29.584 --rc geninfo_unexecuted_blocks=1 00:32:29.584 00:32:29.584 ' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:32:29.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:29.584 --rc genhtml_branch_coverage=1 00:32:29.584 --rc genhtml_function_coverage=1 00:32:29.584 --rc genhtml_legend=1 00:32:29.584 --rc geninfo_all_blocks=1 00:32:29.584 --rc geninfo_unexecuted_blocks=1 00:32:29.584 00:32:29.584 ' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:32:29.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:29.584 --rc genhtml_branch_coverage=1 00:32:29.584 --rc genhtml_function_coverage=1 00:32:29.584 --rc genhtml_legend=1 00:32:29.584 --rc geninfo_all_blocks=1 00:32:29.584 --rc geninfo_unexecuted_blocks=1 00:32:29.584 00:32:29.584 ' 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:32:29.584 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@474 -- # prepare_net_devs 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@436 -- # local -g is_hw=no 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@438 -- # remove_spdk_ns 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:32:29.846 17:53:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:37.985 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:32:37.986 Found 0000:31:00.0 (0x8086 - 0x159b) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:32:37.986 Found 0000:31:00.1 (0x8086 - 0x159b) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:32:37.986 Found net devices under 0000:31:00.0: cvl_0_0 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:32:37.986 Found net devices under 0000:31:00.1: cvl_0_1 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # is_hw=yes 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:37.986 17:53:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:37.986 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:37.986 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.679 ms 00:32:37.986 00:32:37.986 --- 10.0.0.2 ping statistics --- 00:32:37.986 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:37.986 rtt min/avg/max/mdev = 0.679/0.679/0.679/0.000 ms 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:37.986 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:37.986 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.311 ms 00:32:37.986 00:32:37.986 --- 10.0.0.1 ping statistics --- 00:32:37.986 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:37.986 rtt min/avg/max/mdev = 0.311/0.311/0.311/0.000 ms 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@448 -- # return 0 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:32:37.986 only one NIC for nvmf test 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:37.986 rmmod nvme_tcp 00:32:37.986 rmmod nvme_fabrics 00:32:37.986 rmmod nvme_keyring 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:32:37.986 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:32:37.987 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:37.987 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:37.987 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:37.987 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:37.987 17:53:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:39.373 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:39.634 00:32:39.634 real 0m10.030s 00:32:39.634 user 0m2.092s 00:32:39.634 sys 0m5.866s 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:32:39.634 ************************************ 00:32:39.634 END TEST nvmf_target_multipath 00:32:39.634 ************************************ 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:32:39.634 ************************************ 00:32:39.634 START TEST nvmf_zcopy 00:32:39.634 ************************************ 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:32:39.634 * Looking for test storage... 00:32:39.634 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1691 -- # lcov --version 00:32:39.634 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:32:39.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:39.896 --rc genhtml_branch_coverage=1 00:32:39.896 --rc genhtml_function_coverage=1 00:32:39.896 --rc genhtml_legend=1 00:32:39.896 --rc geninfo_all_blocks=1 00:32:39.896 --rc geninfo_unexecuted_blocks=1 00:32:39.896 00:32:39.896 ' 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:32:39.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:39.896 --rc genhtml_branch_coverage=1 00:32:39.896 --rc genhtml_function_coverage=1 00:32:39.896 --rc genhtml_legend=1 00:32:39.896 --rc geninfo_all_blocks=1 00:32:39.896 --rc geninfo_unexecuted_blocks=1 00:32:39.896 00:32:39.896 ' 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:32:39.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:39.896 --rc genhtml_branch_coverage=1 00:32:39.896 --rc genhtml_function_coverage=1 00:32:39.896 --rc genhtml_legend=1 00:32:39.896 --rc geninfo_all_blocks=1 00:32:39.896 --rc geninfo_unexecuted_blocks=1 00:32:39.896 00:32:39.896 ' 00:32:39.896 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:32:39.896 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:39.896 --rc genhtml_branch_coverage=1 00:32:39.896 --rc genhtml_function_coverage=1 00:32:39.897 --rc genhtml_legend=1 00:32:39.897 --rc geninfo_all_blocks=1 00:32:39.897 --rc geninfo_unexecuted_blocks=1 00:32:39.897 00:32:39.897 ' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@474 -- # prepare_net_devs 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@436 -- # local -g is_hw=no 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@438 -- # remove_spdk_ns 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:32:39.897 17:53:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:48.041 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:32:48.042 Found 0000:31:00.0 (0x8086 - 0x159b) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:32:48.042 Found 0000:31:00.1 (0x8086 - 0x159b) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:32:48.042 Found net devices under 0000:31:00.0: cvl_0_0 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:32:48.042 Found net devices under 0000:31:00.1: cvl_0_1 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # is_hw=yes 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:48.042 17:53:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:48.042 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:48.042 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.590 ms 00:32:48.042 00:32:48.042 --- 10.0.0.2 ping statistics --- 00:32:48.042 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:48.042 rtt min/avg/max/mdev = 0.590/0.590/0.590/0.000 ms 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:48.042 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:48.042 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.329 ms 00:32:48.042 00:32:48.042 --- 10.0.0.1 ping statistics --- 00:32:48.042 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:48.042 rtt min/avg/max/mdev = 0.329/0.329/0.329/0.000 ms 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@448 -- # return 0 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@724 -- # xtrace_disable 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@507 -- # nvmfpid=302003 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@508 -- # waitforlisten 302003 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@831 -- # '[' -z 302003 ']' 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:48.042 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:48.043 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:48.043 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:48.043 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:48.043 17:53:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.043 [2024-10-17 17:53:55.334965] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:32:48.043 [2024-10-17 17:53:55.336142] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:48.043 [2024-10-17 17:53:55.336194] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:48.043 [2024-10-17 17:53:55.426935] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:48.043 [2024-10-17 17:53:55.476923] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:48.043 [2024-10-17 17:53:55.476976] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:48.043 [2024-10-17 17:53:55.476984] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:48.043 [2024-10-17 17:53:55.476992] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:48.043 [2024-10-17 17:53:55.476998] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:48.043 [2024-10-17 17:53:55.477852] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:48.043 [2024-10-17 17:53:55.553980] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:32:48.043 [2024-10-17 17:53:55.554281] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@864 -- # return 0 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@730 -- # xtrace_disable 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:48.304 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.304 [2024-10-17 17:53:56.218727] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.566 [2024-10-17 17:53:56.247051] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.566 malloc0 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:48.566 { 00:32:48.566 "params": { 00:32:48.566 "name": "Nvme$subsystem", 00:32:48.566 "trtype": "$TEST_TRANSPORT", 00:32:48.566 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:48.566 "adrfam": "ipv4", 00:32:48.566 "trsvcid": "$NVMF_PORT", 00:32:48.566 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:48.566 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:48.566 "hdgst": ${hdgst:-false}, 00:32:48.566 "ddgst": ${ddgst:-false} 00:32:48.566 }, 00:32:48.566 "method": "bdev_nvme_attach_controller" 00:32:48.566 } 00:32:48.566 EOF 00:32:48.566 )") 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:32:48.566 17:53:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:48.566 "params": { 00:32:48.566 "name": "Nvme1", 00:32:48.566 "trtype": "tcp", 00:32:48.566 "traddr": "10.0.0.2", 00:32:48.566 "adrfam": "ipv4", 00:32:48.566 "trsvcid": "4420", 00:32:48.566 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:48.566 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:48.566 "hdgst": false, 00:32:48.566 "ddgst": false 00:32:48.566 }, 00:32:48.566 "method": "bdev_nvme_attach_controller" 00:32:48.566 }' 00:32:48.566 [2024-10-17 17:53:56.360284] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:48.566 [2024-10-17 17:53:56.360347] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid302127 ] 00:32:48.566 [2024-10-17 17:53:56.439805] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:48.828 [2024-10-17 17:53:56.493025] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:48.828 Running I/O for 10 seconds... 00:32:51.158 6332.00 IOPS, 49.47 MiB/s [2024-10-17T15:54:00.021Z] 6372.50 IOPS, 49.79 MiB/s [2024-10-17T15:54:00.967Z] 6380.33 IOPS, 49.85 MiB/s [2024-10-17T15:54:01.912Z] 6391.00 IOPS, 49.93 MiB/s [2024-10-17T15:54:02.855Z] 6401.20 IOPS, 50.01 MiB/s [2024-10-17T15:54:03.799Z] 6408.67 IOPS, 50.07 MiB/s [2024-10-17T15:54:04.741Z] 6692.43 IOPS, 52.28 MiB/s [2024-10-17T15:54:05.684Z] 7046.75 IOPS, 55.05 MiB/s [2024-10-17T15:54:07.068Z] 7324.56 IOPS, 57.22 MiB/s [2024-10-17T15:54:07.068Z] 7547.20 IOPS, 58.96 MiB/s 00:32:59.149 Latency(us) 00:32:59.149 [2024-10-17T15:54:07.068Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:59.149 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:32:59.149 Verification LBA range: start 0x0 length 0x1000 00:32:59.149 Nvme1n1 : 10.01 7552.22 59.00 0.00 0.00 16901.23 2293.76 28180.48 00:32:59.149 [2024-10-17T15:54:07.068Z] =================================================================================================================== 00:32:59.149 [2024-10-17T15:54:07.068Z] Total : 7552.22 59.00 0.00 0.00 16901.23 2293.76 28180.48 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=304141 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:59.149 { 00:32:59.149 "params": { 00:32:59.149 "name": "Nvme$subsystem", 00:32:59.149 "trtype": "$TEST_TRANSPORT", 00:32:59.149 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:59.149 "adrfam": "ipv4", 00:32:59.149 "trsvcid": "$NVMF_PORT", 00:32:59.149 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:59.149 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:59.149 "hdgst": ${hdgst:-false}, 00:32:59.149 "ddgst": ${ddgst:-false} 00:32:59.149 }, 00:32:59.149 "method": "bdev_nvme_attach_controller" 00:32:59.149 } 00:32:59.149 EOF 00:32:59.149 )") 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:32:59.149 [2024-10-17 17:54:06.798258] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.798286] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:32:59.149 17:54:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:59.149 "params": { 00:32:59.149 "name": "Nvme1", 00:32:59.149 "trtype": "tcp", 00:32:59.149 "traddr": "10.0.0.2", 00:32:59.149 "adrfam": "ipv4", 00:32:59.149 "trsvcid": "4420", 00:32:59.149 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:59.149 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:59.149 "hdgst": false, 00:32:59.149 "ddgst": false 00:32:59.149 }, 00:32:59.149 "method": "bdev_nvme_attach_controller" 00:32:59.149 }' 00:32:59.149 [2024-10-17 17:54:06.810215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.810223] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.822213] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.822221] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.834214] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.834221] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.837783] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:32:59.149 [2024-10-17 17:54:06.837829] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid304141 ] 00:32:59.149 [2024-10-17 17:54:06.846214] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.846221] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.858213] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.858220] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.870213] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.870220] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.882213] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.882220] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.894214] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.894220] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.906213] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.906220] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.916526] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:59.149 [2024-10-17 17:54:06.918214] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.918225] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.930214] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.930222] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.942214] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.942224] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.945794] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:59.149 [2024-10-17 17:54:06.954213] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.954221] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.966219] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.966231] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.978217] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.978228] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:06.990215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:06.990224] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:07.002214] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:07.002222] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:07.014223] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:07.014240] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:07.026216] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:07.026226] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:07.038215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:07.038225] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:07.050215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:07.050226] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.149 [2024-10-17 17:54:07.062215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.149 [2024-10-17 17:54:07.062225] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.074222] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.074239] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 Running I/O for 5 seconds... 00:32:59.411 [2024-10-17 17:54:07.090192] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.090208] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.102215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.102231] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.114439] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.114454] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.127261] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.127277] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.141843] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.141863] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.154686] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.154705] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.169807] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.169822] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.182136] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.182150] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.194635] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.194649] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.209176] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.209191] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.222791] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.222806] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.236571] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.236586] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.249422] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.249437] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.261963] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.261978] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.274942] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.274956] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.288894] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.288909] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.301854] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.301869] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.314486] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.314500] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.411 [2024-10-17 17:54:07.329056] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.411 [2024-10-17 17:54:07.329071] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.342097] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.342112] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.354806] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.354820] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.369276] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.369291] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.381872] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.381887] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.394272] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.394291] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.406913] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.406928] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.421261] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.421276] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.434455] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.434470] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.447523] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.447537] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.460996] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.461011] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.474092] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.474107] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.486341] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.486356] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.499184] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.499199] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.513711] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.513725] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.527254] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.527268] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.542170] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.542184] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.554920] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.554935] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.569727] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.569742] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.672 [2024-10-17 17:54:07.582475] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.672 [2024-10-17 17:54:07.582489] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.597564] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.597579] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.610141] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.610156] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.622252] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.622266] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.635232] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.635246] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.649431] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.649450] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.662242] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.662257] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.674084] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.674099] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.686680] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.686699] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.701219] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.701234] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.714185] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.714200] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.726788] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.726803] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.741418] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.741433] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.753972] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.753987] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.767150] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.767165] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.782002] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.782017] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.794930] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.794944] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.809801] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.809816] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.822305] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.822320] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.833773] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.833788] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:32:59.934 [2024-10-17 17:54:07.846790] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:32:59.934 [2024-10-17 17:54:07.846804] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.861638] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.861653] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.874515] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.874528] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.889435] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.889450] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.902667] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.902687] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.917610] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.917625] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.930493] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.930506] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.945390] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.945405] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.958286] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.958301] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.971277] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.971291] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.985034] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.985048] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:07.997890] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:07.997904] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:08.010208] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.195 [2024-10-17 17:54:08.010223] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.195 [2024-10-17 17:54:08.022855] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.196 [2024-10-17 17:54:08.022869] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.196 [2024-10-17 17:54:08.037320] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.196 [2024-10-17 17:54:08.037335] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.196 [2024-10-17 17:54:08.049905] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.196 [2024-10-17 17:54:08.049920] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.196 [2024-10-17 17:54:08.062393] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.196 [2024-10-17 17:54:08.062406] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.196 [2024-10-17 17:54:08.077326] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.196 [2024-10-17 17:54:08.077341] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.196 18836.00 IOPS, 147.16 MiB/s [2024-10-17T15:54:08.115Z] [2024-10-17 17:54:08.090010] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.196 [2024-10-17 17:54:08.090024] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.196 [2024-10-17 17:54:08.102509] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.196 [2024-10-17 17:54:08.102523] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.117171] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.117185] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.130702] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.130717] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.145354] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.145368] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.158529] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.158543] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.173537] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.173551] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.186250] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.186265] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.197995] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.198010] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.210846] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.210860] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.225238] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.225252] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.238424] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.238439] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.250703] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.250717] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.265471] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.265485] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.278393] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.278408] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.293158] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.293173] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.305799] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.305813] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.318253] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.318268] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.331036] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.331051] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.346004] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.346019] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.358647] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.358660] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.458 [2024-10-17 17:54:08.373456] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.458 [2024-10-17 17:54:08.373471] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.386461] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.386476] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.399314] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.399329] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.413075] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.413089] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.426384] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.426399] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.439293] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.439307] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.453659] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.453674] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.466805] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.466819] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.481192] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.481207] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.494386] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.494401] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.506841] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.506855] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.521401] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.521415] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.534504] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.534518] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.549863] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.549877] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.562686] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.562705] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.577939] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.577954] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.590628] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.590642] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.605744] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.605759] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.618696] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.618710] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.721 [2024-10-17 17:54:08.633437] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.721 [2024-10-17 17:54:08.633452] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.645920] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.645934] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.658601] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.658617] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.673156] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.673171] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.685931] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.685945] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.698227] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.698241] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.711277] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.711291] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.725418] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.725432] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.738204] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.738218] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.750808] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.750822] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.765702] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.765717] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.778546] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.778560] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.793422] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.793437] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.806158] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.806174] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.818159] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.818175] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.831186] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.831200] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.845165] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.845181] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.858132] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.858147] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.870643] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.870657] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.885703] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.885718] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:00.983 [2024-10-17 17:54:08.898756] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:00.983 [2024-10-17 17:54:08.898770] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.244 [2024-10-17 17:54:08.913588] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.244 [2024-10-17 17:54:08.913606] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.244 [2024-10-17 17:54:08.926117] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.244 [2024-10-17 17:54:08.926132] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.244 [2024-10-17 17:54:08.938112] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.244 [2024-10-17 17:54:08.938126] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.244 [2024-10-17 17:54:08.951161] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.244 [2024-10-17 17:54:08.951175] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:08.965182] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:08.965196] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:08.978184] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:08.978199] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:08.990087] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:08.990101] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.002596] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.002610] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.017444] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.017459] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.030390] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.030405] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.042629] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.042643] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.056972] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.056987] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.070220] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.070235] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.082435] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.082450] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 18861.00 IOPS, 147.35 MiB/s [2024-10-17T15:54:09.164Z] [2024-10-17 17:54:09.094712] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.094726] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.109592] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.109607] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.122190] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.122205] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.134709] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.134723] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.149802] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.149817] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.245 [2024-10-17 17:54:09.162476] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.245 [2024-10-17 17:54:09.162494] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.177458] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.177472] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.189845] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.189860] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.202465] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.202480] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.217335] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.217349] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.230400] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.230415] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.242491] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.242505] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.257338] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.257353] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.270291] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.270306] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.282584] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.282598] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.297500] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.297515] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.310703] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.310717] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.325860] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.325874] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.338673] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.338688] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.353469] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.353484] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.366079] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.366093] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.378348] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.378362] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.393712] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.393727] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.407083] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.407099] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.507 [2024-10-17 17:54:09.420819] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.507 [2024-10-17 17:54:09.420837] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.434071] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.434085] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.446649] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.446663] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.461733] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.461748] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.475238] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.475252] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.489301] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.489316] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.502499] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.502513] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.517670] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.517685] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.530418] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.530433] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.542927] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.542941] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.557609] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.557623] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.570231] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.570246] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.581488] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.581502] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.594400] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.594413] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.609297] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.609311] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.622167] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.622181] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.634588] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.634602] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.649567] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.649582] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.662133] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.662147] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:01.769 [2024-10-17 17:54:09.674369] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:01.769 [2024-10-17 17:54:09.674382] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.030 [2024-10-17 17:54:09.689180] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.030 [2024-10-17 17:54:09.689194] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.030 [2024-10-17 17:54:09.702203] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.030 [2024-10-17 17:54:09.702218] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.030 [2024-10-17 17:54:09.714342] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.030 [2024-10-17 17:54:09.714356] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.030 [2024-10-17 17:54:09.729260] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.030 [2024-10-17 17:54:09.729275] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.030 [2024-10-17 17:54:09.742046] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.030 [2024-10-17 17:54:09.742060] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.754369] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.754383] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.766483] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.766496] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.781164] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.781178] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.793882] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.793896] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.806595] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.806609] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.821157] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.821172] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.834362] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.834377] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.846506] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.846520] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.861400] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.861415] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.874226] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.874241] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.887276] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.887290] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.902435] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.902448] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.916786] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.916800] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.929711] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.929725] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.031 [2024-10-17 17:54:09.942540] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.031 [2024-10-17 17:54:09.942554] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:09.957556] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:09.957572] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:09.970355] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:09.970369] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:09.985211] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:09.985226] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:09.998347] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:09.998361] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.010157] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.010175] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.022874] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.022889] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.037371] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.037386] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.050040] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.050056] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.062550] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.062564] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.077486] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.077501] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 18889.67 IOPS, 147.58 MiB/s [2024-10-17T15:54:10.211Z] [2024-10-17 17:54:10.090651] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.090667] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.105036] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.105051] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.117824] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.117839] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.130613] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.130627] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.145498] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.145513] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.158886] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.158901] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.174397] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.174412] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.186846] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.186861] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.292 [2024-10-17 17:54:10.201480] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.292 [2024-10-17 17:54:10.201495] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.214825] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.214839] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.229582] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.229598] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.242283] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.242299] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.254201] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.254216] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.267106] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.267121] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.280993] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.281008] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.294035] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.294050] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.306887] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.306901] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.321853] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.321874] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.334534] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.334548] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.349864] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.349879] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.362745] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.362759] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.377842] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.377857] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.390774] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.390789] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.405572] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.405587] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.418798] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.418812] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.433709] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.433728] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.446609] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.446624] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.554 [2024-10-17 17:54:10.461785] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.554 [2024-10-17 17:54:10.461799] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.475223] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.475239] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.489428] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.489443] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.502317] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.502332] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.514871] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.514885] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.529653] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.529668] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.542329] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.542344] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.555092] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.555106] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.569876] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.569891] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.582752] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.582767] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.597638] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.597653] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.610476] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.610490] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.625545] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.625560] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.638257] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.638273] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.649439] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.649454] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.662369] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.662383] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.676997] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.677012] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.689888] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.689907] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.702451] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.702465] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.717153] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.717169] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.730145] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.730160] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:02.840 [2024-10-17 17:54:10.742519] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:02.840 [2024-10-17 17:54:10.742533] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.757621] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.757637] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.770321] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.770336] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.782340] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.782354] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.797174] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.797189] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.809868] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.809882] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.822679] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.822697] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.837096] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.837111] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.850028] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.850043] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.862537] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.862552] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.877406] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.877421] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.890483] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.890497] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.905745] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.905760] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.918370] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.918385] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.930756] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.930770] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.945205] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.945223] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.957880] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.957895] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.970653] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.970667] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.985730] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.985745] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:10.998524] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:10.998538] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:11.013198] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:11.013212] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.120 [2024-10-17 17:54:11.026179] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.120 [2024-10-17 17:54:11.026193] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.038230] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.038245] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.050980] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.050995] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.065714] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.065728] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.078539] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.078553] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 18904.50 IOPS, 147.69 MiB/s [2024-10-17T15:54:11.300Z] [2024-10-17 17:54:11.093706] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.093721] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.106630] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.106644] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.121694] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.121709] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.134731] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.134745] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.149531] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.149546] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.162749] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.162764] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.177670] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.177685] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.190696] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.190711] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.381 [2024-10-17 17:54:11.205422] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.381 [2024-10-17 17:54:11.205437] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.382 [2024-10-17 17:54:11.218772] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.382 [2024-10-17 17:54:11.218787] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.382 [2024-10-17 17:54:11.233319] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.382 [2024-10-17 17:54:11.233333] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.382 [2024-10-17 17:54:11.246195] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.382 [2024-10-17 17:54:11.246210] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.382 [2024-10-17 17:54:11.258877] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.382 [2024-10-17 17:54:11.258892] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.382 [2024-10-17 17:54:11.273638] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.382 [2024-10-17 17:54:11.273652] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.382 [2024-10-17 17:54:11.286432] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.382 [2024-10-17 17:54:11.286447] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.382 [2024-10-17 17:54:11.298554] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.382 [2024-10-17 17:54:11.298568] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.313128] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.313143] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.326073] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.326087] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.338193] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.338207] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.351190] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.351204] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.365893] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.365908] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.378871] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.378885] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.393721] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.393737] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.406408] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.406421] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.421123] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.421137] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.433967] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.433981] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.446074] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.446089] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.458701] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.458715] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.473809] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.473824] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.486911] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.486926] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.502128] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.502143] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.515027] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.515041] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.529531] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.529545] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.542449] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.542464] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.644 [2024-10-17 17:54:11.555023] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.644 [2024-10-17 17:54:11.555037] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.569890] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.569904] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.582728] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.582742] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.597060] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.597075] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.609833] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.609848] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.622746] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.622760] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.637423] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.637438] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.650537] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.650551] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.665106] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.665120] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.678196] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.678210] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.689931] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.689945] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.702701] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.702716] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.717411] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.717425] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.730523] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.730538] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.745203] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.745218] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.757735] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.757750] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.770745] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.770759] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.785320] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.905 [2024-10-17 17:54:11.785335] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.905 [2024-10-17 17:54:11.798472] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.906 [2024-10-17 17:54:11.798486] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:03.906 [2024-10-17 17:54:11.813363] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:03.906 [2024-10-17 17:54:11.813377] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.826292] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.826307] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.838915] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.838930] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.853197] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.853211] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.866138] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.866152] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.878165] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.878179] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.890905] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.890919] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.904930] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.904945] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.917492] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.917506] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.930135] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.930149] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.942789] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.942803] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.957335] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.957353] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.970552] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.970566] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.985242] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.985256] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:11.998082] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:11.998097] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:12.009934] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:12.009949] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:12.022698] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:12.022712] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:12.037628] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.166 [2024-10-17 17:54:12.037642] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.166 [2024-10-17 17:54:12.050221] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.167 [2024-10-17 17:54:12.050236] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.167 [2024-10-17 17:54:12.062824] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.167 [2024-10-17 17:54:12.062838] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.167 [2024-10-17 17:54:12.076878] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.167 [2024-10-17 17:54:12.076892] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.427 [2024-10-17 17:54:12.089760] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.427 [2024-10-17 17:54:12.089782] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.427 18897.00 IOPS, 147.63 MiB/s 00:33:04.427 Latency(us) 00:33:04.428 [2024-10-17T15:54:12.347Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:04.428 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:33:04.428 Nvme1n1 : 5.00 18904.59 147.69 0.00 0.00 6765.53 2798.93 11414.19 00:33:04.428 [2024-10-17T15:54:12.347Z] =================================================================================================================== 00:33:04.428 [2024-10-17T15:54:12.347Z] Total : 18904.59 147.69 0.00 0.00 6765.53 2798.93 11414.19 00:33:04.428 [2024-10-17 17:54:12.098221] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.098235] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.110220] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.110233] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.122226] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.122238] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.134220] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.134230] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.146218] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.146229] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.158215] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.158231] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.170216] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.170224] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.182219] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.182229] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 [2024-10-17 17:54:12.194224] subsystem.c:2128:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:33:04.428 [2024-10-17 17:54:12.194232] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:04.428 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (304141) - No such process 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 304141 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:33:04.428 delay0 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.428 17:54:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:33:04.428 [2024-10-17 17:54:12.306153] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:33:12.573 Initializing NVMe Controllers 00:33:12.573 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:12.573 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:33:12.573 Initialization complete. Launching workers. 00:33:12.573 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 5069 00:33:12.573 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 5352, failed to submit 37 00:33:12.573 success 5162, unsuccessful 190, failed 0 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@514 -- # nvmfcleanup 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:33:12.573 rmmod nvme_tcp 00:33:12.573 rmmod nvme_fabrics 00:33:12.573 rmmod nvme_keyring 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@515 -- # '[' -n 302003 ']' 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@516 -- # killprocess 302003 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@950 -- # '[' -z 302003 ']' 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@954 -- # kill -0 302003 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@955 -- # uname 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 302003 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@968 -- # echo 'killing process with pid 302003' 00:33:12.573 killing process with pid 302003 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@969 -- # kill 302003 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@974 -- # wait 302003 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-save 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-restore 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:12.573 17:54:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:13.516 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:13.516 00:33:13.516 real 0m33.971s 00:33:13.516 user 0m42.656s 00:33:13.516 sys 0m12.741s 00:33:13.516 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:13.516 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:33:13.516 ************************************ 00:33:13.516 END TEST nvmf_zcopy 00:33:13.516 ************************************ 00:33:13.516 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:33:13.516 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:33:13.516 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:13.516 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:13.778 ************************************ 00:33:13.778 START TEST nvmf_nmic 00:33:13.778 ************************************ 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:33:13.778 * Looking for test storage... 00:33:13.778 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1691 -- # lcov --version 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:33:13.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:13.778 --rc genhtml_branch_coverage=1 00:33:13.778 --rc genhtml_function_coverage=1 00:33:13.778 --rc genhtml_legend=1 00:33:13.778 --rc geninfo_all_blocks=1 00:33:13.778 --rc geninfo_unexecuted_blocks=1 00:33:13.778 00:33:13.778 ' 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:33:13.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:13.778 --rc genhtml_branch_coverage=1 00:33:13.778 --rc genhtml_function_coverage=1 00:33:13.778 --rc genhtml_legend=1 00:33:13.778 --rc geninfo_all_blocks=1 00:33:13.778 --rc geninfo_unexecuted_blocks=1 00:33:13.778 00:33:13.778 ' 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:33:13.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:13.778 --rc genhtml_branch_coverage=1 00:33:13.778 --rc genhtml_function_coverage=1 00:33:13.778 --rc genhtml_legend=1 00:33:13.778 --rc geninfo_all_blocks=1 00:33:13.778 --rc geninfo_unexecuted_blocks=1 00:33:13.778 00:33:13.778 ' 00:33:13.778 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:33:13.778 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:13.778 --rc genhtml_branch_coverage=1 00:33:13.778 --rc genhtml_function_coverage=1 00:33:13.778 --rc genhtml_legend=1 00:33:13.778 --rc geninfo_all_blocks=1 00:33:13.778 --rc geninfo_unexecuted_blocks=1 00:33:13.778 00:33:13.778 ' 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:13.779 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@474 -- # prepare_net_devs 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@436 -- # local -g is_hw=no 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@438 -- # remove_spdk_ns 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:14.040 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:14.041 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:33:14.041 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:33:14.041 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:33:14.041 17:54:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:33:22.190 Found 0000:31:00.0 (0x8086 - 0x159b) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:33:22.190 Found 0000:31:00.1 (0x8086 - 0x159b) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:22.190 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:33:22.191 Found net devices under 0000:31:00.0: cvl_0_0 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:33:22.191 Found net devices under 0000:31:00.1: cvl_0_1 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # is_hw=yes 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:33:22.191 17:54:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:33:22.191 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:33:22.191 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.661 ms 00:33:22.191 00:33:22.191 --- 10.0.0.2 ping statistics --- 00:33:22.191 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:22.191 rtt min/avg/max/mdev = 0.661/0.661/0.661/0.000 ms 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:33:22.191 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:33:22.191 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.305 ms 00:33:22.191 00:33:22.191 --- 10.0.0.1 ping statistics --- 00:33:22.191 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:22.191 rtt min/avg/max/mdev = 0.305/0.305/0.305/0.000 ms 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@448 -- # return 0 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@507 -- # nvmfpid=311182 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@508 -- # waitforlisten 311182 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@831 -- # '[' -z 311182 ']' 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:22.191 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:22.191 17:54:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.191 [2024-10-17 17:54:29.371974] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:33:22.191 [2024-10-17 17:54:29.373103] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:33:22.191 [2024-10-17 17:54:29.373149] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:22.191 [2024-10-17 17:54:29.462185] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:33:22.191 [2024-10-17 17:54:29.517477] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:22.191 [2024-10-17 17:54:29.517530] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:22.191 [2024-10-17 17:54:29.517539] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:22.191 [2024-10-17 17:54:29.517547] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:22.191 [2024-10-17 17:54:29.517554] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:22.191 [2024-10-17 17:54:29.519973] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:33:22.191 [2024-10-17 17:54:29.520134] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:33:22.191 [2024-10-17 17:54:29.520293] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:22.191 [2024-10-17 17:54:29.520293] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:33:22.191 [2024-10-17 17:54:29.597701] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:33:22.191 [2024-10-17 17:54:29.598064] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:33:22.191 [2024-10-17 17:54:29.598708] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:33:22.192 [2024-10-17 17:54:29.599269] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:33:22.192 [2024-10-17 17:54:29.599285] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:33:22.453 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:22.453 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@864 -- # return 0 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.454 [2024-10-17 17:54:30.253179] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.454 Malloc0 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.454 [2024-10-17 17:54:30.345492] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:33:22.454 test case1: single bdev can't be used in multiple subsystems 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.454 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.716 [2024-10-17 17:54:30.380784] bdev.c:8202:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:33:22.716 [2024-10-17 17:54:30.380812] subsystem.c:2157:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:33:22.716 [2024-10-17 17:54:30.380822] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:33:22.716 request: 00:33:22.716 { 00:33:22.716 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:33:22.716 "namespace": { 00:33:22.716 "bdev_name": "Malloc0", 00:33:22.716 "no_auto_visible": false 00:33:22.716 }, 00:33:22.716 "method": "nvmf_subsystem_add_ns", 00:33:22.716 "req_id": 1 00:33:22.716 } 00:33:22.716 Got JSON-RPC error response 00:33:22.716 response: 00:33:22.716 { 00:33:22.716 "code": -32602, 00:33:22.716 "message": "Invalid parameters" 00:33:22.716 } 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:33:22.716 Adding namespace failed - expected result. 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:33:22.716 test case2: host connect to nvmf target in multiple paths 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:22.716 [2024-10-17 17:54:30.392947] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.716 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:33:22.978 17:54:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:33:23.550 17:54:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:33:23.550 17:54:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1198 -- # local i=0 00:33:23.550 17:54:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:33:23.550 17:54:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:33:23.550 17:54:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1205 -- # sleep 2 00:33:25.466 17:54:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:33:25.466 17:54:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:33:25.466 17:54:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:33:25.466 17:54:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:33:25.466 17:54:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:33:25.466 17:54:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1208 -- # return 0 00:33:25.466 17:54:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:33:25.466 [global] 00:33:25.466 thread=1 00:33:25.466 invalidate=1 00:33:25.466 rw=write 00:33:25.466 time_based=1 00:33:25.466 runtime=1 00:33:25.466 ioengine=libaio 00:33:25.466 direct=1 00:33:25.466 bs=4096 00:33:25.466 iodepth=1 00:33:25.466 norandommap=0 00:33:25.466 numjobs=1 00:33:25.466 00:33:25.466 verify_dump=1 00:33:25.466 verify_backlog=512 00:33:25.466 verify_state_save=0 00:33:25.466 do_verify=1 00:33:25.466 verify=crc32c-intel 00:33:25.466 [job0] 00:33:25.466 filename=/dev/nvme0n1 00:33:25.466 Could not set queue depth (nvme0n1) 00:33:25.728 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:25.728 fio-3.35 00:33:25.728 Starting 1 thread 00:33:27.114 00:33:27.114 job0: (groupid=0, jobs=1): err= 0: pid=312169: Thu Oct 17 17:54:34 2024 00:33:27.114 read: IOPS=18, BW=75.5KiB/s (77.4kB/s)(76.0KiB/1006msec) 00:33:27.114 slat (nsec): min=10157, max=26325, avg=24962.63, stdev=3592.00 00:33:27.114 clat (usec): min=40895, max=41202, avg=40979.67, stdev=71.37 00:33:27.114 lat (usec): min=40921, max=41212, avg=41004.63, stdev=68.76 00:33:27.114 clat percentiles (usec): 00:33:27.114 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:33:27.114 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:33:27.114 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:33:27.114 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:33:27.114 | 99.99th=[41157] 00:33:27.114 write: IOPS=508, BW=2036KiB/s (2085kB/s)(2048KiB/1006msec); 0 zone resets 00:33:27.114 slat (nsec): min=9537, max=62698, avg=26264.55, stdev=10665.05 00:33:27.114 clat (usec): min=206, max=593, avg=410.84, stdev=65.80 00:33:27.114 lat (usec): min=238, max=626, avg=437.10, stdev=70.36 00:33:27.114 clat percentiles (usec): 00:33:27.114 | 1.00th=[ 241], 5.00th=[ 302], 10.00th=[ 326], 20.00th=[ 351], 00:33:27.114 | 30.00th=[ 371], 40.00th=[ 400], 50.00th=[ 420], 60.00th=[ 453], 00:33:27.114 | 70.00th=[ 465], 80.00th=[ 469], 90.00th=[ 478], 95.00th=[ 490], 00:33:27.114 | 99.00th=[ 515], 99.50th=[ 537], 99.90th=[ 594], 99.95th=[ 594], 00:33:27.114 | 99.99th=[ 594] 00:33:27.114 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:33:27.114 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:27.114 lat (usec) : 250=1.88%, 500=91.90%, 750=2.64% 00:33:27.114 lat (msec) : 50=3.58% 00:33:27.114 cpu : usr=0.80%, sys=1.19%, ctx=531, majf=0, minf=1 00:33:27.114 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:27.114 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:27.114 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:27.114 issued rwts: total=19,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:27.114 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:27.114 00:33:27.114 Run status group 0 (all jobs): 00:33:27.114 READ: bw=75.5KiB/s (77.4kB/s), 75.5KiB/s-75.5KiB/s (77.4kB/s-77.4kB/s), io=76.0KiB (77.8kB), run=1006-1006msec 00:33:27.114 WRITE: bw=2036KiB/s (2085kB/s), 2036KiB/s-2036KiB/s (2085kB/s-2085kB/s), io=2048KiB (2097kB), run=1006-1006msec 00:33:27.114 00:33:27.114 Disk stats (read/write): 00:33:27.114 nvme0n1: ios=66/512, merge=0/0, ticks=698/210, in_queue=908, util=92.59% 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:33:27.114 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1219 -- # local i=0 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1231 -- # return 0 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@514 -- # nvmfcleanup 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:33:27.114 rmmod nvme_tcp 00:33:27.114 rmmod nvme_fabrics 00:33:27.114 rmmod nvme_keyring 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@515 -- # '[' -n 311182 ']' 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@516 -- # killprocess 311182 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@950 -- # '[' -z 311182 ']' 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@954 -- # kill -0 311182 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@955 -- # uname 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:27.114 17:54:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 311182 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@968 -- # echo 'killing process with pid 311182' 00:33:27.375 killing process with pid 311182 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@969 -- # kill 311182 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@974 -- # wait 311182 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-save 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-restore 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:27.375 17:54:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:29.920 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:29.920 00:33:29.920 real 0m15.787s 00:33:29.920 user 0m35.616s 00:33:29.920 sys 0m7.623s 00:33:29.920 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:29.920 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:33:29.920 ************************************ 00:33:29.921 END TEST nvmf_nmic 00:33:29.921 ************************************ 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:29.921 ************************************ 00:33:29.921 START TEST nvmf_fio_target 00:33:29.921 ************************************ 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:33:29.921 * Looking for test storage... 00:33:29.921 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1691 -- # lcov --version 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:33:29.921 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:29.921 --rc genhtml_branch_coverage=1 00:33:29.921 --rc genhtml_function_coverage=1 00:33:29.921 --rc genhtml_legend=1 00:33:29.921 --rc geninfo_all_blocks=1 00:33:29.921 --rc geninfo_unexecuted_blocks=1 00:33:29.921 00:33:29.921 ' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:33:29.921 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:29.921 --rc genhtml_branch_coverage=1 00:33:29.921 --rc genhtml_function_coverage=1 00:33:29.921 --rc genhtml_legend=1 00:33:29.921 --rc geninfo_all_blocks=1 00:33:29.921 --rc geninfo_unexecuted_blocks=1 00:33:29.921 00:33:29.921 ' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:33:29.921 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:29.921 --rc genhtml_branch_coverage=1 00:33:29.921 --rc genhtml_function_coverage=1 00:33:29.921 --rc genhtml_legend=1 00:33:29.921 --rc geninfo_all_blocks=1 00:33:29.921 --rc geninfo_unexecuted_blocks=1 00:33:29.921 00:33:29.921 ' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:33:29.921 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:29.921 --rc genhtml_branch_coverage=1 00:33:29.921 --rc genhtml_function_coverage=1 00:33:29.921 --rc genhtml_legend=1 00:33:29.921 --rc geninfo_all_blocks=1 00:33:29.921 --rc geninfo_unexecuted_blocks=1 00:33:29.921 00:33:29.921 ' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:33:29.921 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:33:29.922 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:33:38.063 Found 0000:31:00.0 (0x8086 - 0x159b) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:33:38.063 Found 0000:31:00.1 (0x8086 - 0x159b) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:33:38.063 Found net devices under 0000:31:00.0: cvl_0_0 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:33:38.063 Found net devices under 0000:31:00.1: cvl_0_1 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # is_hw=yes 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:33:38.063 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:33:38.063 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:33:38.063 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:33:38.063 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:33:38.063 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:33:38.063 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:33:38.063 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.650 ms 00:33:38.063 00:33:38.063 --- 10.0.0.2 ping statistics --- 00:33:38.063 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:38.063 rtt min/avg/max/mdev = 0.650/0.650/0.650/0.000 ms 00:33:38.063 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:33:38.063 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:33:38.063 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.289 ms 00:33:38.063 00:33:38.063 --- 10.0.0.1 ping statistics --- 00:33:38.063 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:38.063 rtt min/avg/max/mdev = 0.289/0.289/0.289/0.000 ms 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@448 -- # return 0 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@507 -- # nvmfpid=316538 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@508 -- # waitforlisten 316538 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@831 -- # '[' -z 316538 ']' 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:38.064 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:38.064 17:54:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:33:38.064 [2024-10-17 17:54:45.209992] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:33:38.064 [2024-10-17 17:54:45.211139] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:33:38.064 [2024-10-17 17:54:45.211189] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:38.064 [2024-10-17 17:54:45.300187] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:33:38.064 [2024-10-17 17:54:45.354632] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:38.064 [2024-10-17 17:54:45.354679] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:38.064 [2024-10-17 17:54:45.354687] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:38.064 [2024-10-17 17:54:45.354705] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:38.064 [2024-10-17 17:54:45.354711] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:38.064 [2024-10-17 17:54:45.356746] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:33:38.064 [2024-10-17 17:54:45.356960] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:33:38.064 [2024-10-17 17:54:45.357128] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:33:38.064 [2024-10-17 17:54:45.357129] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:38.064 [2024-10-17 17:54:45.435370] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:33:38.064 [2024-10-17 17:54:45.436545] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:33:38.064 [2024-10-17 17:54:45.436623] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:33:38.064 [2024-10-17 17:54:45.437046] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:33:38.064 [2024-10-17 17:54:45.437142] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:33:38.325 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:38.325 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@864 -- # return 0 00:33:38.325 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:33:38.325 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:38.325 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:33:38.325 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:38.325 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:33:38.325 [2024-10-17 17:54:46.234180] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:38.585 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:33:38.585 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:33:38.585 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:33:38.846 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:33:38.846 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:33:39.106 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:33:39.106 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:33:39.367 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:33:39.367 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:33:39.628 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:33:39.628 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:33:39.628 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:33:39.888 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:33:39.888 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:33:40.148 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:33:40.148 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:33:40.409 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:33:40.409 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:33:40.409 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:33:40.670 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:33:40.670 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:33:40.931 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:33:40.931 [2024-10-17 17:54:48.826116] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:41.193 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:33:41.193 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:33:41.455 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:33:42.028 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:33:42.028 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1198 -- # local i=0 00:33:42.028 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:33:42.028 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1200 -- # [[ -n 4 ]] 00:33:42.028 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1201 -- # nvme_device_counter=4 00:33:42.028 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1205 -- # sleep 2 00:33:43.944 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:33:43.944 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:33:43.944 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:33:43.944 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1207 -- # nvme_devices=4 00:33:43.944 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:33:43.944 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1208 -- # return 0 00:33:43.944 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:33:43.944 [global] 00:33:43.944 thread=1 00:33:43.944 invalidate=1 00:33:43.944 rw=write 00:33:43.944 time_based=1 00:33:43.944 runtime=1 00:33:43.944 ioengine=libaio 00:33:43.944 direct=1 00:33:43.944 bs=4096 00:33:43.944 iodepth=1 00:33:43.944 norandommap=0 00:33:43.944 numjobs=1 00:33:43.944 00:33:43.944 verify_dump=1 00:33:43.944 verify_backlog=512 00:33:43.944 verify_state_save=0 00:33:43.944 do_verify=1 00:33:43.944 verify=crc32c-intel 00:33:43.944 [job0] 00:33:43.944 filename=/dev/nvme0n1 00:33:43.944 [job1] 00:33:43.944 filename=/dev/nvme0n2 00:33:43.944 [job2] 00:33:43.944 filename=/dev/nvme0n3 00:33:43.944 [job3] 00:33:43.944 filename=/dev/nvme0n4 00:33:43.944 Could not set queue depth (nvme0n1) 00:33:43.944 Could not set queue depth (nvme0n2) 00:33:43.944 Could not set queue depth (nvme0n3) 00:33:43.944 Could not set queue depth (nvme0n4) 00:33:44.206 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:44.206 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:44.206 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:44.206 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:44.206 fio-3.35 00:33:44.206 Starting 4 threads 00:33:45.592 00:33:45.592 job0: (groupid=0, jobs=1): err= 0: pid=318114: Thu Oct 17 17:54:53 2024 00:33:45.592 read: IOPS=16, BW=67.7KiB/s (69.4kB/s)(68.0KiB/1004msec) 00:33:45.592 slat (nsec): min=24807, max=25846, avg=25108.88, stdev=251.68 00:33:45.592 clat (usec): min=959, max=42070, avg=39293.31, stdev=9887.48 00:33:45.592 lat (usec): min=984, max=42095, avg=39318.42, stdev=9887.43 00:33:45.592 clat percentiles (usec): 00:33:45.592 | 1.00th=[ 963], 5.00th=[ 963], 10.00th=[41157], 20.00th=[41157], 00:33:45.592 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[42206], 00:33:45.592 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:33:45.592 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:33:45.592 | 99.99th=[42206] 00:33:45.592 write: IOPS=509, BW=2040KiB/s (2089kB/s)(2048KiB/1004msec); 0 zone resets 00:33:45.592 slat (nsec): min=9821, max=78304, avg=31948.74, stdev=6574.68 00:33:45.592 clat (usec): min=269, max=1112, avg=616.14, stdev=149.71 00:33:45.592 lat (usec): min=280, max=1144, avg=648.09, stdev=150.77 00:33:45.592 clat percentiles (usec): 00:33:45.592 | 1.00th=[ 314], 5.00th=[ 388], 10.00th=[ 416], 20.00th=[ 469], 00:33:45.592 | 30.00th=[ 529], 40.00th=[ 578], 50.00th=[ 627], 60.00th=[ 660], 00:33:45.592 | 70.00th=[ 701], 80.00th=[ 750], 90.00th=[ 799], 95.00th=[ 865], 00:33:45.592 | 99.00th=[ 947], 99.50th=[ 1020], 99.90th=[ 1106], 99.95th=[ 1106], 00:33:45.592 | 99.99th=[ 1106] 00:33:45.592 bw ( KiB/s): min= 4096, max= 4096, per=45.22%, avg=4096.00, stdev= 0.00, samples=1 00:33:45.592 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:45.592 lat (usec) : 500=23.82%, 750=55.39%, 1000=17.01% 00:33:45.592 lat (msec) : 2=0.76%, 50=3.02% 00:33:45.592 cpu : usr=0.90%, sys=1.50%, ctx=530, majf=0, minf=1 00:33:45.592 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:45.592 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.592 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.592 issued rwts: total=17,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:45.592 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:45.592 job1: (groupid=0, jobs=1): err= 0: pid=318115: Thu Oct 17 17:54:53 2024 00:33:45.592 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:33:45.592 slat (nsec): min=8300, max=61957, avg=26148.49, stdev=4762.53 00:33:45.592 clat (usec): min=679, max=1312, avg=990.75, stdev=100.61 00:33:45.593 lat (usec): min=688, max=1338, avg=1016.90, stdev=101.90 00:33:45.593 clat percentiles (usec): 00:33:45.593 | 1.00th=[ 742], 5.00th=[ 791], 10.00th=[ 848], 20.00th=[ 914], 00:33:45.593 | 30.00th=[ 947], 40.00th=[ 988], 50.00th=[ 1004], 60.00th=[ 1029], 00:33:45.593 | 70.00th=[ 1045], 80.00th=[ 1057], 90.00th=[ 1106], 95.00th=[ 1139], 00:33:45.593 | 99.00th=[ 1221], 99.50th=[ 1237], 99.90th=[ 1319], 99.95th=[ 1319], 00:33:45.593 | 99.99th=[ 1319] 00:33:45.593 write: IOPS=741, BW=2965KiB/s (3036kB/s)(2968KiB/1001msec); 0 zone resets 00:33:45.593 slat (usec): min=9, max=2105, avg=35.45, stdev=87.34 00:33:45.593 clat (usec): min=257, max=963, avg=594.04, stdev=117.41 00:33:45.593 lat (usec): min=268, max=2684, avg=629.49, stdev=153.17 00:33:45.593 clat percentiles (usec): 00:33:45.593 | 1.00th=[ 330], 5.00th=[ 392], 10.00th=[ 445], 20.00th=[ 490], 00:33:45.593 | 30.00th=[ 529], 40.00th=[ 570], 50.00th=[ 603], 60.00th=[ 627], 00:33:45.593 | 70.00th=[ 652], 80.00th=[ 693], 90.00th=[ 742], 95.00th=[ 783], 00:33:45.593 | 99.00th=[ 865], 99.50th=[ 906], 99.90th=[ 963], 99.95th=[ 963], 00:33:45.593 | 99.99th=[ 963] 00:33:45.593 bw ( KiB/s): min= 4096, max= 4096, per=45.22%, avg=4096.00, stdev= 0.00, samples=1 00:33:45.593 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:45.593 lat (usec) : 500=12.84%, 750=41.07%, 1000=24.32% 00:33:45.593 lat (msec) : 2=21.77% 00:33:45.593 cpu : usr=2.20%, sys=3.30%, ctx=1257, majf=0, minf=1 00:33:45.593 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:45.593 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.593 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.593 issued rwts: total=512,742,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:45.593 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:45.593 job2: (groupid=0, jobs=1): err= 0: pid=318116: Thu Oct 17 17:54:53 2024 00:33:45.593 read: IOPS=14, BW=59.9KiB/s (61.4kB/s)(60.0KiB/1001msec) 00:33:45.593 slat (nsec): min=10782, max=26870, avg=25306.13, stdev=4025.05 00:33:45.593 clat (usec): min=41017, max=42096, avg=41864.94, stdev=269.03 00:33:45.593 lat (usec): min=41043, max=42123, avg=41890.25, stdev=270.47 00:33:45.593 clat percentiles (usec): 00:33:45.593 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41681], 20.00th=[41681], 00:33:45.593 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[42206], 00:33:45.593 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:33:45.593 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:33:45.593 | 99.99th=[42206] 00:33:45.593 write: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec); 0 zone resets 00:33:45.593 slat (usec): min=9, max=13093, avg=57.20, stdev=577.33 00:33:45.593 clat (usec): min=248, max=3315, avg=657.17, stdev=182.75 00:33:45.593 lat (usec): min=259, max=13974, avg=714.37, stdev=615.67 00:33:45.593 clat percentiles (usec): 00:33:45.593 | 1.00th=[ 334], 5.00th=[ 408], 10.00th=[ 465], 20.00th=[ 515], 00:33:45.593 | 30.00th=[ 586], 40.00th=[ 627], 50.00th=[ 668], 60.00th=[ 701], 00:33:45.593 | 70.00th=[ 734], 80.00th=[ 766], 90.00th=[ 799], 95.00th=[ 881], 00:33:45.593 | 99.00th=[ 963], 99.50th=[ 1012], 99.90th=[ 3326], 99.95th=[ 3326], 00:33:45.593 | 99.99th=[ 3326] 00:33:45.593 bw ( KiB/s): min= 4096, max= 4096, per=45.22%, avg=4096.00, stdev= 0.00, samples=1 00:33:45.593 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:45.593 lat (usec) : 250=0.19%, 500=15.94%, 750=56.93%, 1000=23.53% 00:33:45.593 lat (msec) : 2=0.38%, 4=0.19%, 50=2.85% 00:33:45.593 cpu : usr=0.80%, sys=1.60%, ctx=530, majf=0, minf=1 00:33:45.593 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:45.593 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.593 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.593 issued rwts: total=15,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:45.593 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:45.593 job3: (groupid=0, jobs=1): err= 0: pid=318117: Thu Oct 17 17:54:53 2024 00:33:45.593 read: IOPS=16, BW=67.6KiB/s (69.2kB/s)(68.0KiB/1006msec) 00:33:45.593 slat (nsec): min=26514, max=27984, avg=27251.94, stdev=452.92 00:33:45.593 clat (usec): min=40754, max=42065, avg=41492.94, stdev=530.10 00:33:45.593 lat (usec): min=40782, max=42091, avg=41520.19, stdev=529.92 00:33:45.593 clat percentiles (usec): 00:33:45.593 | 1.00th=[40633], 5.00th=[40633], 10.00th=[40633], 20.00th=[41157], 00:33:45.593 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41681], 60.00th=[41681], 00:33:45.593 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:33:45.593 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:33:45.593 | 99.99th=[42206] 00:33:45.593 write: IOPS=508, BW=2036KiB/s (2085kB/s)(2048KiB/1006msec); 0 zone resets 00:33:45.593 slat (nsec): min=9181, max=68303, avg=32832.89, stdev=8403.66 00:33:45.593 clat (usec): min=234, max=1004, avg=546.15, stdev=150.37 00:33:45.593 lat (usec): min=245, max=1038, avg=578.98, stdev=153.13 00:33:45.593 clat percentiles (usec): 00:33:45.593 | 1.00th=[ 251], 5.00th=[ 302], 10.00th=[ 343], 20.00th=[ 400], 00:33:45.593 | 30.00th=[ 457], 40.00th=[ 502], 50.00th=[ 553], 60.00th=[ 603], 00:33:45.593 | 70.00th=[ 644], 80.00th=[ 676], 90.00th=[ 725], 95.00th=[ 766], 00:33:45.593 | 99.00th=[ 881], 99.50th=[ 914], 99.90th=[ 1004], 99.95th=[ 1004], 00:33:45.593 | 99.99th=[ 1004] 00:33:45.593 bw ( KiB/s): min= 4096, max= 4096, per=45.22%, avg=4096.00, stdev= 0.00, samples=1 00:33:45.593 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:45.593 lat (usec) : 250=0.95%, 500=36.48%, 750=52.17%, 1000=6.99% 00:33:45.593 lat (msec) : 2=0.19%, 50=3.21% 00:33:45.593 cpu : usr=1.29%, sys=1.99%, ctx=529, majf=0, minf=1 00:33:45.593 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:45.593 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.593 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:45.593 issued rwts: total=17,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:45.593 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:45.593 00:33:45.593 Run status group 0 (all jobs): 00:33:45.593 READ: bw=2231KiB/s (2284kB/s), 59.9KiB/s-2046KiB/s (61.4kB/s-2095kB/s), io=2244KiB (2298kB), run=1001-1006msec 00:33:45.593 WRITE: bw=9058KiB/s (9275kB/s), 2036KiB/s-2965KiB/s (2085kB/s-3036kB/s), io=9112KiB (9331kB), run=1001-1006msec 00:33:45.593 00:33:45.593 Disk stats (read/write): 00:33:45.593 nvme0n1: ios=62/512, merge=0/0, ticks=517/294, in_queue=811, util=83.77% 00:33:45.593 nvme0n2: ios=504/512, merge=0/0, ticks=647/297, in_queue=944, util=91.01% 00:33:45.593 nvme0n3: ios=32/512, merge=0/0, ticks=1294/328, in_queue=1622, util=95.30% 00:33:45.593 nvme0n4: ios=68/512, merge=0/0, ticks=580/200, in_queue=780, util=96.66% 00:33:45.593 17:54:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:33:45.593 [global] 00:33:45.593 thread=1 00:33:45.593 invalidate=1 00:33:45.593 rw=randwrite 00:33:45.593 time_based=1 00:33:45.593 runtime=1 00:33:45.593 ioengine=libaio 00:33:45.593 direct=1 00:33:45.593 bs=4096 00:33:45.593 iodepth=1 00:33:45.593 norandommap=0 00:33:45.593 numjobs=1 00:33:45.593 00:33:45.593 verify_dump=1 00:33:45.593 verify_backlog=512 00:33:45.593 verify_state_save=0 00:33:45.593 do_verify=1 00:33:45.593 verify=crc32c-intel 00:33:45.593 [job0] 00:33:45.593 filename=/dev/nvme0n1 00:33:45.593 [job1] 00:33:45.593 filename=/dev/nvme0n2 00:33:45.593 [job2] 00:33:45.593 filename=/dev/nvme0n3 00:33:45.593 [job3] 00:33:45.593 filename=/dev/nvme0n4 00:33:45.593 Could not set queue depth (nvme0n1) 00:33:45.593 Could not set queue depth (nvme0n2) 00:33:45.593 Could not set queue depth (nvme0n3) 00:33:45.593 Could not set queue depth (nvme0n4) 00:33:46.163 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:46.163 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:46.163 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:46.163 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:46.163 fio-3.35 00:33:46.163 Starting 4 threads 00:33:47.106 00:33:47.106 job0: (groupid=0, jobs=1): err= 0: pid=318636: Thu Oct 17 17:54:55 2024 00:33:47.106 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:33:47.106 slat (nsec): min=27480, max=50448, avg=28488.79, stdev=2792.14 00:33:47.106 clat (usec): min=660, max=1333, avg=1020.53, stdev=93.38 00:33:47.106 lat (usec): min=688, max=1361, avg=1049.02, stdev=93.02 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 766], 5.00th=[ 848], 10.00th=[ 898], 20.00th=[ 955], 00:33:47.106 | 30.00th=[ 988], 40.00th=[ 1012], 50.00th=[ 1029], 60.00th=[ 1045], 00:33:47.106 | 70.00th=[ 1074], 80.00th=[ 1090], 90.00th=[ 1123], 95.00th=[ 1156], 00:33:47.106 | 99.00th=[ 1237], 99.50th=[ 1270], 99.90th=[ 1336], 99.95th=[ 1336], 00:33:47.106 | 99.99th=[ 1336] 00:33:47.106 write: IOPS=693, BW=2773KiB/s (2840kB/s)(2776KiB/1001msec); 0 zone resets 00:33:47.106 slat (nsec): min=9118, max=53851, avg=31374.74, stdev=9484.45 00:33:47.106 clat (usec): min=214, max=1019, avg=616.78, stdev=120.67 00:33:47.106 lat (usec): min=224, max=1053, avg=648.15, stdev=124.77 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 314], 5.00th=[ 412], 10.00th=[ 453], 20.00th=[ 515], 00:33:47.106 | 30.00th=[ 562], 40.00th=[ 594], 50.00th=[ 627], 60.00th=[ 660], 00:33:47.106 | 70.00th=[ 693], 80.00th=[ 717], 90.00th=[ 766], 95.00th=[ 791], 00:33:47.106 | 99.00th=[ 881], 99.50th=[ 889], 99.90th=[ 1020], 99.95th=[ 1020], 00:33:47.106 | 99.99th=[ 1020] 00:33:47.106 bw ( KiB/s): min= 4096, max= 4096, per=31.66%, avg=4096.00, stdev= 0.00, samples=1 00:33:47.106 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:47.106 lat (usec) : 250=0.25%, 500=9.87%, 750=40.80%, 1000=21.48% 00:33:47.106 lat (msec) : 2=27.61% 00:33:47.106 cpu : usr=3.30%, sys=4.10%, ctx=1207, majf=0, minf=1 00:33:47.106 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:47.106 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 issued rwts: total=512,694,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:47.106 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:47.106 job1: (groupid=0, jobs=1): err= 0: pid=318637: Thu Oct 17 17:54:55 2024 00:33:47.106 read: IOPS=16, BW=67.6KiB/s (69.2kB/s)(68.0KiB/1006msec) 00:33:47.106 slat (nsec): min=17146, max=30181, avg=23877.29, stdev=4093.03 00:33:47.106 clat (usec): min=1069, max=42164, avg=39541.98, stdev=9914.86 00:33:47.106 lat (usec): min=1095, max=42181, avg=39565.86, stdev=9914.25 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 1074], 5.00th=[ 1074], 10.00th=[41681], 20.00th=[41681], 00:33:47.106 | 30.00th=[41681], 40.00th=[41681], 50.00th=[42206], 60.00th=[42206], 00:33:47.106 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:33:47.106 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:33:47.106 | 99.99th=[42206] 00:33:47.106 write: IOPS=508, BW=2036KiB/s (2085kB/s)(2048KiB/1006msec); 0 zone resets 00:33:47.106 slat (nsec): min=2935, max=57832, avg=27435.38, stdev=10336.79 00:33:47.106 clat (usec): min=255, max=965, avg=609.35, stdev=128.25 00:33:47.106 lat (usec): min=267, max=999, avg=636.78, stdev=132.15 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 355], 5.00th=[ 392], 10.00th=[ 445], 20.00th=[ 490], 00:33:47.106 | 30.00th=[ 529], 40.00th=[ 578], 50.00th=[ 619], 60.00th=[ 644], 00:33:47.106 | 70.00th=[ 685], 80.00th=[ 725], 90.00th=[ 775], 95.00th=[ 807], 00:33:47.106 | 99.00th=[ 914], 99.50th=[ 963], 99.90th=[ 963], 99.95th=[ 963], 00:33:47.106 | 99.99th=[ 963] 00:33:47.106 bw ( KiB/s): min= 4087, max= 4087, per=31.59%, avg=4087.00, stdev= 0.00, samples=1 00:33:47.106 iops : min= 1021, max= 1021, avg=1021.00, stdev= 0.00, samples=1 00:33:47.106 lat (usec) : 500=21.74%, 750=61.63%, 1000=13.42% 00:33:47.106 lat (msec) : 2=0.19%, 50=3.02% 00:33:47.106 cpu : usr=0.70%, sys=1.29%, ctx=531, majf=0, minf=1 00:33:47.106 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:47.106 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 issued rwts: total=17,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:47.106 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:47.106 job2: (groupid=0, jobs=1): err= 0: pid=318638: Thu Oct 17 17:54:55 2024 00:33:47.106 read: IOPS=679, BW=2717KiB/s (2782kB/s)(2720KiB/1001msec) 00:33:47.106 slat (nsec): min=6547, max=75706, avg=25105.76, stdev=7101.97 00:33:47.106 clat (usec): min=276, max=1109, avg=778.48, stdev=113.47 00:33:47.106 lat (usec): min=284, max=1133, avg=803.58, stdev=115.13 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 420], 5.00th=[ 594], 10.00th=[ 627], 20.00th=[ 685], 00:33:47.106 | 30.00th=[ 734], 40.00th=[ 758], 50.00th=[ 791], 60.00th=[ 824], 00:33:47.106 | 70.00th=[ 848], 80.00th=[ 873], 90.00th=[ 906], 95.00th=[ 930], 00:33:47.106 | 99.00th=[ 1004], 99.50th=[ 1057], 99.90th=[ 1106], 99.95th=[ 1106], 00:33:47.106 | 99.99th=[ 1106] 00:33:47.106 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:33:47.106 slat (nsec): min=8791, max=55966, avg=29335.04, stdev=9367.77 00:33:47.106 clat (usec): min=128, max=787, avg=401.64, stdev=114.41 00:33:47.106 lat (usec): min=148, max=823, avg=430.98, stdev=116.38 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 198], 5.00th=[ 231], 10.00th=[ 269], 20.00th=[ 302], 00:33:47.106 | 30.00th=[ 322], 40.00th=[ 351], 50.00th=[ 392], 60.00th=[ 433], 00:33:47.106 | 70.00th=[ 465], 80.00th=[ 502], 90.00th=[ 562], 95.00th=[ 611], 00:33:47.106 | 99.00th=[ 676], 99.50th=[ 701], 99.90th=[ 734], 99.95th=[ 791], 00:33:47.106 | 99.99th=[ 791] 00:33:47.106 bw ( KiB/s): min= 4096, max= 4096, per=31.66%, avg=4096.00, stdev= 0.00, samples=1 00:33:47.106 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:47.106 lat (usec) : 250=4.58%, 500=44.01%, 750=26.12%, 1000=24.88% 00:33:47.106 lat (msec) : 2=0.41% 00:33:47.106 cpu : usr=3.10%, sys=6.60%, ctx=1704, majf=0, minf=2 00:33:47.106 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:47.106 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 issued rwts: total=680,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:47.106 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:47.106 job3: (groupid=0, jobs=1): err= 0: pid=318639: Thu Oct 17 17:54:55 2024 00:33:47.106 read: IOPS=666, BW=2665KiB/s (2729kB/s)(2668KiB/1001msec) 00:33:47.106 slat (nsec): min=6601, max=46995, avg=26585.88, stdev=6925.73 00:33:47.106 clat (usec): min=346, max=1076, avg=774.30, stdev=111.03 00:33:47.106 lat (usec): min=375, max=1105, avg=800.88, stdev=113.16 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 474], 5.00th=[ 586], 10.00th=[ 627], 20.00th=[ 676], 00:33:47.106 | 30.00th=[ 725], 40.00th=[ 758], 50.00th=[ 783], 60.00th=[ 816], 00:33:47.106 | 70.00th=[ 848], 80.00th=[ 873], 90.00th=[ 906], 95.00th=[ 930], 00:33:47.106 | 99.00th=[ 979], 99.50th=[ 988], 99.90th=[ 1074], 99.95th=[ 1074], 00:33:47.106 | 99.99th=[ 1074] 00:33:47.106 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:33:47.106 slat (nsec): min=9041, max=67252, avg=31301.99, stdev=10212.13 00:33:47.106 clat (usec): min=117, max=723, avg=407.99, stdev=116.64 00:33:47.106 lat (usec): min=127, max=758, avg=439.29, stdev=119.13 00:33:47.106 clat percentiles (usec): 00:33:47.106 | 1.00th=[ 176], 5.00th=[ 229], 10.00th=[ 269], 20.00th=[ 310], 00:33:47.106 | 30.00th=[ 330], 40.00th=[ 355], 50.00th=[ 404], 60.00th=[ 445], 00:33:47.106 | 70.00th=[ 474], 80.00th=[ 519], 90.00th=[ 570], 95.00th=[ 603], 00:33:47.106 | 99.00th=[ 668], 99.50th=[ 685], 99.90th=[ 701], 99.95th=[ 725], 00:33:47.106 | 99.99th=[ 725] 00:33:47.106 bw ( KiB/s): min= 4096, max= 4096, per=31.66%, avg=4096.00, stdev= 0.00, samples=1 00:33:47.106 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:33:47.106 lat (usec) : 250=5.20%, 500=41.81%, 750=28.62%, 1000=24.31% 00:33:47.106 lat (msec) : 2=0.06% 00:33:47.106 cpu : usr=4.00%, sys=6.10%, ctx=1692, majf=0, minf=1 00:33:47.106 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:47.106 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:47.106 issued rwts: total=667,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:47.106 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:47.106 00:33:47.106 Run status group 0 (all jobs): 00:33:47.106 READ: bw=7459KiB/s (7638kB/s), 67.6KiB/s-2717KiB/s (69.2kB/s-2782kB/s), io=7504KiB (7684kB), run=1001-1006msec 00:33:47.106 WRITE: bw=12.6MiB/s (13.2MB/s), 2036KiB/s-4092KiB/s (2085kB/s-4190kB/s), io=12.7MiB (13.3MB), run=1001-1006msec 00:33:47.106 00:33:47.106 Disk stats (read/write): 00:33:47.106 nvme0n1: ios=526/512, merge=0/0, ticks=584/256, in_queue=840, util=85.17% 00:33:47.106 nvme0n2: ios=64/512, merge=0/0, ticks=637/297, in_queue=934, util=89.32% 00:33:47.106 nvme0n3: ios=569/979, merge=0/0, ticks=419/306, in_queue=725, util=94.75% 00:33:47.106 nvme0n4: ios=573/956, merge=0/0, ticks=680/317, in_queue=997, util=94.49% 00:33:47.368 17:54:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:33:47.368 [global] 00:33:47.368 thread=1 00:33:47.368 invalidate=1 00:33:47.368 rw=write 00:33:47.368 time_based=1 00:33:47.368 runtime=1 00:33:47.368 ioengine=libaio 00:33:47.368 direct=1 00:33:47.368 bs=4096 00:33:47.368 iodepth=128 00:33:47.368 norandommap=0 00:33:47.368 numjobs=1 00:33:47.368 00:33:47.368 verify_dump=1 00:33:47.368 verify_backlog=512 00:33:47.368 verify_state_save=0 00:33:47.368 do_verify=1 00:33:47.368 verify=crc32c-intel 00:33:47.368 [job0] 00:33:47.368 filename=/dev/nvme0n1 00:33:47.368 [job1] 00:33:47.368 filename=/dev/nvme0n2 00:33:47.368 [job2] 00:33:47.368 filename=/dev/nvme0n3 00:33:47.368 [job3] 00:33:47.368 filename=/dev/nvme0n4 00:33:47.368 Could not set queue depth (nvme0n1) 00:33:47.368 Could not set queue depth (nvme0n2) 00:33:47.368 Could not set queue depth (nvme0n3) 00:33:47.368 Could not set queue depth (nvme0n4) 00:33:47.629 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:47.629 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:47.629 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:47.629 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:47.629 fio-3.35 00:33:47.629 Starting 4 threads 00:33:49.017 00:33:49.017 job0: (groupid=0, jobs=1): err= 0: pid=319159: Thu Oct 17 17:54:56 2024 00:33:49.017 read: IOPS=3555, BW=13.9MiB/s (14.6MB/s)(14.0MiB/1008msec) 00:33:49.017 slat (nsec): min=877, max=11216k, avg=95752.96, stdev=597913.94 00:33:49.017 clat (usec): min=6527, max=40639, avg=11797.00, stdev=4623.19 00:33:49.017 lat (usec): min=6531, max=43386, avg=11892.75, stdev=4683.08 00:33:49.017 clat percentiles (usec): 00:33:49.017 | 1.00th=[ 7242], 5.00th=[ 8586], 10.00th=[ 9110], 20.00th=[ 9503], 00:33:49.017 | 30.00th=[ 9634], 40.00th=[ 9765], 50.00th=[10028], 60.00th=[10290], 00:33:49.017 | 70.00th=[10683], 80.00th=[13173], 90.00th=[18482], 95.00th=[21365], 00:33:49.018 | 99.00th=[30016], 99.50th=[33817], 99.90th=[40633], 99.95th=[40633], 00:33:49.018 | 99.99th=[40633] 00:33:49.018 write: IOPS=3737, BW=14.6MiB/s (15.3MB/s)(14.7MiB/1008msec); 0 zone resets 00:33:49.018 slat (nsec): min=1582, max=15669k, avg=169336.44, stdev=842234.83 00:33:49.018 clat (usec): min=6730, max=72261, avg=22661.15, stdev=17536.96 00:33:49.018 lat (usec): min=6734, max=72270, avg=22830.49, stdev=17649.72 00:33:49.018 clat percentiles (usec): 00:33:49.018 | 1.00th=[ 8094], 5.00th=[ 8717], 10.00th=[ 9241], 20.00th=[ 9503], 00:33:49.018 | 30.00th=[11469], 40.00th=[13698], 50.00th=[13960], 60.00th=[18744], 00:33:49.018 | 70.00th=[24511], 80.00th=[33817], 90.00th=[58459], 95.00th=[65274], 00:33:49.018 | 99.00th=[69731], 99.50th=[71828], 99.90th=[71828], 99.95th=[71828], 00:33:49.018 | 99.99th=[71828] 00:33:49.018 bw ( KiB/s): min=12736, max=16384, per=14.68%, avg=14560.00, stdev=2579.53, samples=2 00:33:49.018 iops : min= 3184, max= 4096, avg=3640.00, stdev=644.88, samples=2 00:33:49.018 lat (msec) : 10=37.04%, 20=40.21%, 50=16.68%, 100=6.07% 00:33:49.018 cpu : usr=2.48%, sys=3.18%, ctx=416, majf=0, minf=1 00:33:49.018 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.1% 00:33:49.018 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:49.018 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:49.018 issued rwts: total=3584,3767,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:49.018 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:49.018 job1: (groupid=0, jobs=1): err= 0: pid=319160: Thu Oct 17 17:54:56 2024 00:33:49.018 read: IOPS=5503, BW=21.5MiB/s (22.5MB/s)(21.6MiB/1006msec) 00:33:49.018 slat (nsec): min=926, max=8960.9k, avg=82190.71, stdev=551475.69 00:33:49.018 clat (usec): min=2784, max=32472, avg=9635.35, stdev=3879.19 00:33:49.018 lat (usec): min=2787, max=32480, avg=9717.54, stdev=3918.41 00:33:49.018 clat percentiles (usec): 00:33:49.018 | 1.00th=[ 5080], 5.00th=[ 5866], 10.00th=[ 6390], 20.00th=[ 7439], 00:33:49.018 | 30.00th=[ 7767], 40.00th=[ 8094], 50.00th=[ 8455], 60.00th=[ 9503], 00:33:49.018 | 70.00th=[10159], 80.00th=[10683], 90.00th=[13304], 95.00th=[17433], 00:33:49.018 | 99.00th=[27919], 99.50th=[29230], 99.90th=[30802], 99.95th=[32375], 00:33:49.018 | 99.99th=[32375] 00:33:49.018 write: IOPS=5598, BW=21.9MiB/s (22.9MB/s)(22.0MiB/1006msec); 0 zone resets 00:33:49.018 slat (nsec): min=1608, max=8121.1k, avg=91185.02, stdev=458003.09 00:33:49.018 clat (usec): min=1193, max=32494, avg=13177.00, stdev=6834.22 00:33:49.018 lat (usec): min=1205, max=32519, avg=13268.18, stdev=6881.74 00:33:49.018 clat percentiles (usec): 00:33:49.018 | 1.00th=[ 3654], 5.00th=[ 4948], 10.00th=[ 5211], 20.00th=[ 6915], 00:33:49.018 | 30.00th=[ 7832], 40.00th=[ 9634], 50.00th=[10421], 60.00th=[13829], 00:33:49.018 | 70.00th=[16188], 80.00th=[20841], 90.00th=[23987], 95.00th=[25297], 00:33:49.018 | 99.00th=[27132], 99.50th=[27132], 99.90th=[27657], 99.95th=[32375], 00:33:49.018 | 99.99th=[32375] 00:33:49.018 bw ( KiB/s): min=20480, max=24576, per=22.71%, avg=22528.00, stdev=2896.31, samples=2 00:33:49.018 iops : min= 5120, max= 6144, avg=5632.00, stdev=724.08, samples=2 00:33:49.018 lat (msec) : 2=0.02%, 4=0.89%, 10=55.07%, 20=30.56%, 50=13.47% 00:33:49.018 cpu : usr=4.48%, sys=5.77%, ctx=522, majf=0, minf=2 00:33:49.018 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:33:49.018 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:49.018 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:49.018 issued rwts: total=5537,5632,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:49.018 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:49.018 job2: (groupid=0, jobs=1): err= 0: pid=319161: Thu Oct 17 17:54:56 2024 00:33:49.018 read: IOPS=7634, BW=29.8MiB/s (31.3MB/s)(30.0MiB/1006msec) 00:33:49.018 slat (nsec): min=1019, max=8061.9k, avg=64273.24, stdev=481797.75 00:33:49.018 clat (usec): min=3033, max=15965, avg=8581.48, stdev=1961.48 00:33:49.018 lat (usec): min=3040, max=15971, avg=8645.75, stdev=1988.55 00:33:49.018 clat percentiles (usec): 00:33:49.018 | 1.00th=[ 4817], 5.00th=[ 5669], 10.00th=[ 6390], 20.00th=[ 7177], 00:33:49.018 | 30.00th=[ 7635], 40.00th=[ 7963], 50.00th=[ 8160], 60.00th=[ 8455], 00:33:49.018 | 70.00th=[ 8979], 80.00th=[10159], 90.00th=[11469], 95.00th=[12125], 00:33:49.018 | 99.00th=[14353], 99.50th=[14615], 99.90th=[15401], 99.95th=[15533], 00:33:49.018 | 99.99th=[15926] 00:33:49.018 write: IOPS=7867, BW=30.7MiB/s (32.2MB/s)(30.9MiB/1006msec); 0 zone resets 00:33:49.018 slat (nsec): min=1748, max=6906.4k, avg=58774.63, stdev=421165.89 00:33:49.018 clat (usec): min=1204, max=15411, avg=7781.54, stdev=1719.52 00:33:49.018 lat (usec): min=1215, max=15415, avg=7840.31, stdev=1731.97 00:33:49.018 clat percentiles (usec): 00:33:49.018 | 1.00th=[ 4424], 5.00th=[ 5211], 10.00th=[ 5473], 20.00th=[ 6259], 00:33:49.018 | 30.00th=[ 6980], 40.00th=[ 7504], 50.00th=[ 7898], 60.00th=[ 8160], 00:33:49.018 | 70.00th=[ 8291], 80.00th=[ 8586], 90.00th=[10421], 95.00th=[10814], 00:33:49.018 | 99.00th=[11863], 99.50th=[12387], 99.90th=[14746], 99.95th=[15008], 00:33:49.018 | 99.99th=[15401] 00:33:49.018 bw ( KiB/s): min=29544, max=32752, per=31.40%, avg=31148.00, stdev=2268.40, samples=2 00:33:49.018 iops : min= 7386, max= 8188, avg=7787.00, stdev=567.10, samples=2 00:33:49.018 lat (msec) : 2=0.02%, 4=0.33%, 10=81.84%, 20=17.81% 00:33:49.018 cpu : usr=5.67%, sys=9.05%, ctx=505, majf=0, minf=1 00:33:49.018 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:33:49.018 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:49.018 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:49.018 issued rwts: total=7680,7915,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:49.018 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:49.018 job3: (groupid=0, jobs=1): err= 0: pid=319162: Thu Oct 17 17:54:56 2024 00:33:49.018 read: IOPS=7563, BW=29.5MiB/s (31.0MB/s)(29.8MiB/1008msec) 00:33:49.018 slat (nsec): min=964, max=8227.0k, avg=65765.94, stdev=483056.45 00:33:49.018 clat (usec): min=3202, max=18381, avg=8786.98, stdev=2123.12 00:33:49.018 lat (usec): min=3208, max=18386, avg=8852.75, stdev=2147.04 00:33:49.018 clat percentiles (usec): 00:33:49.018 | 1.00th=[ 5342], 5.00th=[ 5997], 10.00th=[ 6325], 20.00th=[ 7046], 00:33:49.018 | 30.00th=[ 7570], 40.00th=[ 8029], 50.00th=[ 8356], 60.00th=[ 8848], 00:33:49.018 | 70.00th=[ 9503], 80.00th=[10552], 90.00th=[11994], 95.00th=[12911], 00:33:49.018 | 99.00th=[14484], 99.50th=[15008], 99.90th=[15664], 99.95th=[16450], 00:33:49.018 | 99.99th=[18482] 00:33:49.018 write: IOPS=7619, BW=29.8MiB/s (31.2MB/s)(30.0MiB/1008msec); 0 zone resets 00:33:49.018 slat (nsec): min=1634, max=6649.2k, avg=59625.05, stdev=404297.60 00:33:49.018 clat (usec): min=1243, max=15643, avg=7920.79, stdev=1795.65 00:33:49.018 lat (usec): min=1254, max=15645, avg=7980.41, stdev=1804.96 00:33:49.018 clat percentiles (usec): 00:33:49.018 | 1.00th=[ 3621], 5.00th=[ 5080], 10.00th=[ 5669], 20.00th=[ 6325], 00:33:49.018 | 30.00th=[ 7046], 40.00th=[ 7570], 50.00th=[ 8029], 60.00th=[ 8291], 00:33:49.018 | 70.00th=[ 8455], 80.00th=[ 8848], 90.00th=[10683], 95.00th=[11076], 00:33:49.018 | 99.00th=[11600], 99.50th=[11863], 99.90th=[15139], 99.95th=[15533], 00:33:49.018 | 99.99th=[15664] 00:33:49.018 bw ( KiB/s): min=29456, max=32048, per=31.01%, avg=30752.00, stdev=1832.82, samples=2 00:33:49.018 iops : min= 7364, max= 8012, avg=7688.00, stdev=458.21, samples=2 00:33:49.018 lat (msec) : 2=0.01%, 4=0.88%, 10=78.57%, 20=20.54% 00:33:49.018 cpu : usr=5.26%, sys=8.04%, ctx=566, majf=0, minf=1 00:33:49.018 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:33:49.018 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:49.018 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:49.018 issued rwts: total=7624,7680,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:49.018 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:49.018 00:33:49.018 Run status group 0 (all jobs): 00:33:49.018 READ: bw=94.7MiB/s (99.2MB/s), 13.9MiB/s-29.8MiB/s (14.6MB/s-31.3MB/s), io=95.4MiB (100MB), run=1006-1008msec 00:33:49.018 WRITE: bw=96.9MiB/s (102MB/s), 14.6MiB/s-30.7MiB/s (15.3MB/s-32.2MB/s), io=97.6MiB (102MB), run=1006-1008msec 00:33:49.018 00:33:49.018 Disk stats (read/write): 00:33:49.018 nvme0n1: ios=3077/3072, merge=0/0, ticks=17048/34713, in_queue=51761, util=88.48% 00:33:49.018 nvme0n2: ios=4658/4782, merge=0/0, ticks=41809/60353, in_queue=102162, util=92.47% 00:33:49.018 nvme0n3: ios=6340/6656, merge=0/0, ticks=52214/50137, in_queue=102351, util=97.17% 00:33:49.018 nvme0n4: ios=6218/6656, merge=0/0, ticks=51258/50567, in_queue=101825, util=96.82% 00:33:49.018 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:33:49.018 [global] 00:33:49.018 thread=1 00:33:49.018 invalidate=1 00:33:49.018 rw=randwrite 00:33:49.018 time_based=1 00:33:49.018 runtime=1 00:33:49.018 ioengine=libaio 00:33:49.018 direct=1 00:33:49.018 bs=4096 00:33:49.018 iodepth=128 00:33:49.018 norandommap=0 00:33:49.018 numjobs=1 00:33:49.018 00:33:49.018 verify_dump=1 00:33:49.018 verify_backlog=512 00:33:49.018 verify_state_save=0 00:33:49.018 do_verify=1 00:33:49.018 verify=crc32c-intel 00:33:49.018 [job0] 00:33:49.018 filename=/dev/nvme0n1 00:33:49.018 [job1] 00:33:49.018 filename=/dev/nvme0n2 00:33:49.018 [job2] 00:33:49.018 filename=/dev/nvme0n3 00:33:49.018 [job3] 00:33:49.018 filename=/dev/nvme0n4 00:33:49.018 Could not set queue depth (nvme0n1) 00:33:49.018 Could not set queue depth (nvme0n2) 00:33:49.018 Could not set queue depth (nvme0n3) 00:33:49.018 Could not set queue depth (nvme0n4) 00:33:49.278 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:49.278 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:49.278 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:49.278 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:33:49.278 fio-3.35 00:33:49.278 Starting 4 threads 00:33:50.665 00:33:50.665 job0: (groupid=0, jobs=1): err= 0: pid=319674: Thu Oct 17 17:54:58 2024 00:33:50.665 read: IOPS=5350, BW=20.9MiB/s (21.9MB/s)(21.0MiB/1004msec) 00:33:50.665 slat (nsec): min=966, max=10482k, avg=100920.18, stdev=664497.67 00:33:50.665 clat (usec): min=2514, max=31577, avg=11799.03, stdev=5201.03 00:33:50.665 lat (usec): min=2518, max=31586, avg=11899.95, stdev=5239.96 00:33:50.665 clat percentiles (usec): 00:33:50.665 | 1.00th=[ 3130], 5.00th=[ 5997], 10.00th=[ 6390], 20.00th=[ 7111], 00:33:50.665 | 30.00th=[ 8979], 40.00th=[10159], 50.00th=[10945], 60.00th=[11207], 00:33:50.665 | 70.00th=[12780], 80.00th=[14746], 90.00th=[19530], 95.00th=[22676], 00:33:50.665 | 99.00th=[27919], 99.50th=[29754], 99.90th=[31065], 99.95th=[31589], 00:33:50.665 | 99.99th=[31589] 00:33:50.665 write: IOPS=5609, BW=21.9MiB/s (23.0MB/s)(22.0MiB/1004msec); 0 zone resets 00:33:50.665 slat (nsec): min=1565, max=10324k, avg=76571.89, stdev=354715.29 00:33:50.665 clat (usec): min=1762, max=31540, avg=11353.63, stdev=5098.84 00:33:50.665 lat (usec): min=1770, max=31542, avg=11430.20, stdev=5134.48 00:33:50.665 clat percentiles (usec): 00:33:50.665 | 1.00th=[ 2638], 5.00th=[ 4555], 10.00th=[ 4883], 20.00th=[ 8094], 00:33:50.665 | 30.00th=[10028], 40.00th=[10814], 50.00th=[11207], 60.00th=[11338], 00:33:50.665 | 70.00th=[11469], 80.00th=[11600], 90.00th=[19268], 95.00th=[22414], 00:33:50.665 | 99.00th=[30016], 99.50th=[31065], 99.90th=[31065], 99.95th=[31065], 00:33:50.665 | 99.99th=[31589] 00:33:50.665 bw ( KiB/s): min=20439, max=24576, per=27.53%, avg=22507.50, stdev=2925.30, samples=2 00:33:50.665 iops : min= 5109, max= 6144, avg=5626.50, stdev=731.86, samples=2 00:33:50.665 lat (msec) : 2=0.05%, 4=2.33%, 10=30.23%, 20=58.22%, 50=9.18% 00:33:50.665 cpu : usr=2.99%, sys=5.88%, ctx=685, majf=0, minf=1 00:33:50.665 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:33:50.665 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:50.665 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:50.665 issued rwts: total=5372,5632,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:50.665 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:50.666 job1: (groupid=0, jobs=1): err= 0: pid=319683: Thu Oct 17 17:54:58 2024 00:33:50.666 read: IOPS=5079, BW=19.8MiB/s (20.8MB/s)(20.0MiB/1008msec) 00:33:50.666 slat (nsec): min=900, max=10591k, avg=93697.39, stdev=660117.05 00:33:50.666 clat (usec): min=3028, max=26932, avg=11226.90, stdev=3072.29 00:33:50.666 lat (usec): min=3034, max=26934, avg=11320.60, stdev=3118.54 00:33:50.666 clat percentiles (usec): 00:33:50.666 | 1.00th=[ 4948], 5.00th=[ 6390], 10.00th=[ 6783], 20.00th=[ 9110], 00:33:50.666 | 30.00th=[10028], 40.00th=[10814], 50.00th=[11076], 60.00th=[11207], 00:33:50.666 | 70.00th=[12518], 80.00th=[13435], 90.00th=[14746], 95.00th=[15795], 00:33:50.666 | 99.00th=[20317], 99.50th=[23725], 99.90th=[26346], 99.95th=[26870], 00:33:50.666 | 99.99th=[26870] 00:33:50.666 write: IOPS=5323, BW=20.8MiB/s (21.8MB/s)(21.0MiB/1008msec); 0 zone resets 00:33:50.666 slat (nsec): min=1555, max=8460.1k, avg=92543.23, stdev=399496.45 00:33:50.666 clat (usec): min=1083, max=55536, avg=13110.64, stdev=6899.43 00:33:50.666 lat (usec): min=1092, max=55544, avg=13203.18, stdev=6938.00 00:33:50.666 clat percentiles (usec): 00:33:50.666 | 1.00th=[ 4293], 5.00th=[ 6718], 10.00th=[ 8160], 20.00th=[10552], 00:33:50.666 | 30.00th=[11076], 40.00th=[11207], 50.00th=[11338], 60.00th=[11469], 00:33:50.666 | 70.00th=[11600], 80.00th=[14484], 90.00th=[21627], 95.00th=[24249], 00:33:50.666 | 99.00th=[50070], 99.50th=[52691], 99.90th=[55313], 99.95th=[55313], 00:33:50.666 | 99.99th=[55313] 00:33:50.666 bw ( KiB/s): min=20480, max=21432, per=25.63%, avg=20956.00, stdev=673.17, samples=2 00:33:50.666 iops : min= 5120, max= 5358, avg=5239.00, stdev=168.29, samples=2 00:33:50.666 lat (msec) : 2=0.05%, 4=0.47%, 10=23.08%, 20=69.90%, 50=6.00% 00:33:50.666 lat (msec) : 100=0.51% 00:33:50.666 cpu : usr=3.28%, sys=4.87%, ctx=686, majf=0, minf=2 00:33:50.666 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:33:50.666 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:50.666 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:50.666 issued rwts: total=5120,5366,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:50.666 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:50.666 job2: (groupid=0, jobs=1): err= 0: pid=319684: Thu Oct 17 17:54:58 2024 00:33:50.666 read: IOPS=4316, BW=16.9MiB/s (17.7MB/s)(17.7MiB/1047msec) 00:33:50.666 slat (nsec): min=969, max=7631.3k, avg=91016.18, stdev=527548.30 00:33:50.666 clat (usec): min=6046, max=60083, avg=12892.30, stdev=7462.19 00:33:50.666 lat (usec): min=6049, max=60089, avg=12983.32, stdev=7481.94 00:33:50.666 clat percentiles (usec): 00:33:50.666 | 1.00th=[ 6783], 5.00th=[ 7635], 10.00th=[ 8455], 20.00th=[ 8979], 00:33:50.666 | 30.00th=[ 9372], 40.00th=[10814], 50.00th=[12649], 60.00th=[12911], 00:33:50.666 | 70.00th=[13173], 80.00th=[14353], 90.00th=[15926], 95.00th=[16581], 00:33:50.666 | 99.00th=[53216], 99.50th=[56886], 99.90th=[60031], 99.95th=[60031], 00:33:50.666 | 99.99th=[60031] 00:33:50.666 write: IOPS=4401, BW=17.2MiB/s (18.0MB/s)(18.0MiB/1047msec); 0 zone resets 00:33:50.666 slat (nsec): min=1607, max=20938k, avg=122565.77, stdev=738899.18 00:33:50.666 clat (usec): min=4803, max=65811, avg=15640.28, stdev=8286.10 00:33:50.666 lat (usec): min=4813, max=65844, avg=15762.84, stdev=8360.58 00:33:50.666 clat percentiles (usec): 00:33:50.666 | 1.00th=[ 6915], 5.00th=[ 7832], 10.00th=[ 8094], 20.00th=[10421], 00:33:50.666 | 30.00th=[11469], 40.00th=[12518], 50.00th=[13042], 60.00th=[13566], 00:33:50.666 | 70.00th=[15926], 80.00th=[19530], 90.00th=[25560], 95.00th=[32113], 00:33:50.666 | 99.00th=[49546], 99.50th=[53216], 99.90th=[53216], 99.95th=[53740], 00:33:50.666 | 99.99th=[65799] 00:33:50.666 bw ( KiB/s): min=16440, max=20424, per=22.54%, avg=18432.00, stdev=2817.11, samples=2 00:33:50.666 iops : min= 4110, max= 5106, avg=4608.00, stdev=704.28, samples=2 00:33:50.666 lat (msec) : 10=26.30%, 20=62.91%, 50=9.21%, 100=1.58% 00:33:50.666 cpu : usr=2.77%, sys=4.68%, ctx=468, majf=0, minf=1 00:33:50.666 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:33:50.666 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:50.666 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:50.666 issued rwts: total=4519,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:50.666 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:50.666 job3: (groupid=0, jobs=1): err= 0: pid=319685: Thu Oct 17 17:54:58 2024 00:33:50.666 read: IOPS=5598, BW=21.9MiB/s (22.9MB/s)(22.0MiB/1006msec) 00:33:50.666 slat (nsec): min=989, max=6138.3k, avg=80626.01, stdev=486086.77 00:33:50.666 clat (usec): min=2577, max=26238, avg=10417.15, stdev=3682.69 00:33:50.666 lat (usec): min=2583, max=26248, avg=10497.77, stdev=3723.54 00:33:50.666 clat percentiles (usec): 00:33:50.666 | 1.00th=[ 4490], 5.00th=[ 5276], 10.00th=[ 6259], 20.00th=[ 7439], 00:33:50.666 | 30.00th=[ 8225], 40.00th=[ 8848], 50.00th=[ 9503], 60.00th=[10552], 00:33:50.666 | 70.00th=[12649], 80.00th=[13173], 90.00th=[15926], 95.00th=[17171], 00:33:50.666 | 99.00th=[20055], 99.50th=[22938], 99.90th=[26084], 99.95th=[26346], 00:33:50.666 | 99.99th=[26346] 00:33:50.666 write: IOPS=5761, BW=22.5MiB/s (23.6MB/s)(22.6MiB/1006msec); 0 zone resets 00:33:50.666 slat (nsec): min=1624, max=6364.8k, avg=87313.21, stdev=443356.69 00:33:50.666 clat (usec): min=772, max=36508, avg=11862.58, stdev=6666.89 00:33:50.666 lat (usec): min=1547, max=36515, avg=11949.89, stdev=6713.89 00:33:50.666 clat percentiles (usec): 00:33:50.666 | 1.00th=[ 3294], 5.00th=[ 4359], 10.00th=[ 5800], 20.00th=[ 6587], 00:33:50.666 | 30.00th=[ 7373], 40.00th=[ 8586], 50.00th=[10421], 60.00th=[11994], 00:33:50.666 | 70.00th=[13042], 80.00th=[15270], 90.00th=[23462], 95.00th=[26608], 00:33:50.666 | 99.00th=[30802], 99.50th=[32637], 99.90th=[36439], 99.95th=[36439], 00:33:50.666 | 99.99th=[36439] 00:33:50.666 bw ( KiB/s): min=17896, max=27448, per=27.73%, avg=22672.00, stdev=6754.28, samples=2 00:33:50.666 iops : min= 4474, max= 6862, avg=5668.00, stdev=1688.57, samples=2 00:33:50.666 lat (usec) : 1000=0.01% 00:33:50.666 lat (msec) : 2=0.10%, 4=1.61%, 10=50.78%, 20=39.68%, 50=7.82% 00:33:50.666 cpu : usr=3.78%, sys=5.87%, ctx=589, majf=0, minf=1 00:33:50.666 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:33:50.666 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:50.666 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:33:50.666 issued rwts: total=5632,5796,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:50.666 latency : target=0, window=0, percentile=100.00%, depth=128 00:33:50.666 00:33:50.666 Run status group 0 (all jobs): 00:33:50.666 READ: bw=77.0MiB/s (80.8MB/s), 16.9MiB/s-21.9MiB/s (17.7MB/s-22.9MB/s), io=80.6MiB (84.6MB), run=1004-1047msec 00:33:50.666 WRITE: bw=79.8MiB/s (83.7MB/s), 17.2MiB/s-22.5MiB/s (18.0MB/s-23.6MB/s), io=83.6MiB (87.7MB), run=1004-1047msec 00:33:50.666 00:33:50.666 Disk stats (read/write): 00:33:50.666 nvme0n1: ios=4146/4607, merge=0/0, ticks=50034/52277, in_queue=102311, util=88.18% 00:33:50.666 nvme0n2: ios=4381/4608, merge=0/0, ticks=49363/54287, in_queue=103650, util=87.27% 00:33:50.666 nvme0n3: ios=3617/3759, merge=0/0, ticks=19527/27346, in_queue=46873, util=99.16% 00:33:50.666 nvme0n4: ios=4966/5120, merge=0/0, ticks=30468/33330, in_queue=63798, util=100.00% 00:33:50.666 17:54:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:33:50.666 17:54:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=319783 00:33:50.666 17:54:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:33:50.666 17:54:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:33:50.666 [global] 00:33:50.666 thread=1 00:33:50.666 invalidate=1 00:33:50.666 rw=read 00:33:50.666 time_based=1 00:33:50.666 runtime=10 00:33:50.666 ioengine=libaio 00:33:50.666 direct=1 00:33:50.666 bs=4096 00:33:50.666 iodepth=1 00:33:50.666 norandommap=1 00:33:50.666 numjobs=1 00:33:50.666 00:33:50.666 [job0] 00:33:50.666 filename=/dev/nvme0n1 00:33:50.666 [job1] 00:33:50.666 filename=/dev/nvme0n2 00:33:50.666 [job2] 00:33:50.666 filename=/dev/nvme0n3 00:33:50.666 [job3] 00:33:50.666 filename=/dev/nvme0n4 00:33:50.666 Could not set queue depth (nvme0n1) 00:33:50.666 Could not set queue depth (nvme0n2) 00:33:50.666 Could not set queue depth (nvme0n3) 00:33:50.666 Could not set queue depth (nvme0n4) 00:33:50.928 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:50.928 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:50.928 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:50.928 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:33:50.928 fio-3.35 00:33:50.928 Starting 4 threads 00:33:53.471 17:55:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:33:53.731 17:55:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:33:53.731 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=262144, buflen=4096 00:33:53.731 fio: pid=320171, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:33:53.992 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=1949696, buflen=4096 00:33:53.992 fio: pid=320165, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:33:53.992 17:55:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:33:53.992 17:55:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:33:54.255 17:55:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:33:54.255 17:55:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:33:54.255 fio: io_u error on file /dev/nvme0n1: Operation not supported: read offset=1617920, buflen=4096 00:33:54.255 fio: pid=320134, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:33:54.255 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:33:54.255 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:33:54.255 fio: io_u error on file /dev/nvme0n2: Operation not supported: read offset=434176, buflen=4096 00:33:54.255 fio: pid=320149, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:33:54.255 00:33:54.255 job0: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=320134: Thu Oct 17 17:55:02 2024 00:33:54.255 read: IOPS=132, BW=529KiB/s (542kB/s)(1580KiB/2986msec) 00:33:54.255 slat (usec): min=7, max=213, avg=25.76, stdev=10.55 00:33:54.255 clat (usec): min=470, max=42485, avg=7473.12, stdev=15067.35 00:33:54.255 lat (usec): min=495, max=42510, avg=7498.88, stdev=15068.75 00:33:54.255 clat percentiles (usec): 00:33:54.255 | 1.00th=[ 553], 5.00th=[ 578], 10.00th=[ 660], 20.00th=[ 750], 00:33:54.255 | 30.00th=[ 824], 40.00th=[ 881], 50.00th=[ 930], 60.00th=[ 947], 00:33:54.255 | 70.00th=[ 971], 80.00th=[ 1037], 90.00th=[41681], 95.00th=[42206], 00:33:54.255 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42730], 99.95th=[42730], 00:33:54.255 | 99.99th=[42730] 00:33:54.255 bw ( KiB/s): min= 88, max= 2672, per=46.52%, avg=612.60, stdev=1151.27, samples=5 00:33:54.255 iops : min= 22, max= 668, avg=153.00, stdev=287.90, samples=5 00:33:54.255 lat (usec) : 500=0.25%, 750=18.69%, 1000=58.59% 00:33:54.255 lat (msec) : 2=6.06%, 50=16.16% 00:33:54.255 cpu : usr=0.13%, sys=0.40%, ctx=398, majf=0, minf=1 00:33:54.255 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:54.255 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.255 complete : 0=0.3%, 4=99.7%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.255 issued rwts: total=396,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:54.255 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:54.255 job1: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=320149: Thu Oct 17 17:55:02 2024 00:33:54.255 read: IOPS=33, BW=134KiB/s (137kB/s)(424KiB/3165msec) 00:33:54.255 slat (usec): min=10, max=7596, avg=99.91, stdev=731.58 00:33:54.255 clat (usec): min=526, max=42079, avg=29544.95, stdev=18568.97 00:33:54.255 lat (usec): min=552, max=48897, avg=29645.50, stdev=18628.95 00:33:54.255 clat percentiles (usec): 00:33:54.255 | 1.00th=[ 603], 5.00th=[ 799], 10.00th=[ 865], 20.00th=[ 1012], 00:33:54.255 | 30.00th=[23462], 40.00th=[41157], 50.00th=[41157], 60.00th=[41681], 00:33:54.255 | 70.00th=[41681], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:33:54.255 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:33:54.255 | 99.99th=[42206] 00:33:54.255 bw ( KiB/s): min= 95, max= 296, per=10.26%, avg=135.83, stdev=80.10, samples=6 00:33:54.255 iops : min= 23, max= 74, avg=33.83, stdev=20.10, samples=6 00:33:54.255 lat (usec) : 750=3.74%, 1000=15.89% 00:33:54.255 lat (msec) : 2=9.35%, 50=70.09% 00:33:54.255 cpu : usr=0.00%, sys=0.19%, ctx=111, majf=0, minf=2 00:33:54.255 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:54.255 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.255 complete : 0=0.9%, 4=99.1%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.255 issued rwts: total=107,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:54.255 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:54.255 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=320165: Thu Oct 17 17:55:02 2024 00:33:54.255 read: IOPS=171, BW=686KiB/s (702kB/s)(1904KiB/2777msec) 00:33:54.255 slat (usec): min=3, max=5365, avg=38.43, stdev=244.49 00:33:54.255 clat (usec): min=700, max=42163, avg=5742.06, stdev=12791.92 00:33:54.255 lat (usec): min=727, max=42196, avg=5780.51, stdev=12790.21 00:33:54.255 clat percentiles (usec): 00:33:54.255 | 1.00th=[ 865], 5.00th=[ 963], 10.00th=[ 1020], 20.00th=[ 1074], 00:33:54.255 | 30.00th=[ 1106], 40.00th=[ 1123], 50.00th=[ 1139], 60.00th=[ 1156], 00:33:54.255 | 70.00th=[ 1188], 80.00th=[ 1221], 90.00th=[41157], 95.00th=[41157], 00:33:54.255 | 99.00th=[41157], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:33:54.255 | 99.99th=[42206] 00:33:54.255 bw ( KiB/s): min= 96, max= 1493, per=50.85%, avg=669.80, stdev=609.28, samples=5 00:33:54.255 iops : min= 24, max= 373, avg=167.40, stdev=152.24, samples=5 00:33:54.255 lat (usec) : 750=0.21%, 1000=7.55% 00:33:54.255 lat (msec) : 2=80.08%, 4=0.42%, 50=11.53% 00:33:54.255 cpu : usr=0.18%, sys=0.79%, ctx=479, majf=0, minf=1 00:33:54.255 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:54.255 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.255 complete : 0=0.2%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.255 issued rwts: total=477,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:54.255 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:54.255 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=320171: Thu Oct 17 17:55:02 2024 00:33:54.255 read: IOPS=24, BW=98.0KiB/s (100kB/s)(256KiB/2611msec) 00:33:54.255 slat (nsec): min=25675, max=41718, avg=26503.18, stdev=1942.49 00:33:54.255 clat (usec): min=906, max=41959, avg=40359.74, stdev=5012.51 00:33:54.255 lat (usec): min=948, max=41985, avg=40386.22, stdev=5010.58 00:33:54.255 clat percentiles (usec): 00:33:54.255 | 1.00th=[ 906], 5.00th=[40633], 10.00th=[40633], 20.00th=[40633], 00:33:54.255 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:33:54.255 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:33:54.255 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:33:54.255 | 99.99th=[42206] 00:33:54.255 bw ( KiB/s): min= 96, max= 104, per=7.52%, avg=99.00, stdev= 4.12, samples=5 00:33:54.255 iops : min= 24, max= 26, avg=24.60, stdev= 0.89, samples=5 00:33:54.255 lat (usec) : 1000=1.54% 00:33:54.255 lat (msec) : 50=96.92% 00:33:54.256 cpu : usr=0.11%, sys=0.00%, ctx=66, majf=0, minf=2 00:33:54.256 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:33:54.256 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.256 complete : 0=1.5%, 4=98.5%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:33:54.256 issued rwts: total=65,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:33:54.256 latency : target=0, window=0, percentile=100.00%, depth=1 00:33:54.256 00:33:54.256 Run status group 0 (all jobs): 00:33:54.256 READ: bw=1316KiB/s (1347kB/s), 98.0KiB/s-686KiB/s (100kB/s-702kB/s), io=4164KiB (4264kB), run=2611-3165msec 00:33:54.256 00:33:54.256 Disk stats (read/write): 00:33:54.256 nvme0n1: ios=392/0, merge=0/0, ticks=2825/0, in_queue=2825, util=94.99% 00:33:54.256 nvme0n2: ios=132/0, merge=0/0, ticks=3758/0, in_queue=3758, util=99.88% 00:33:54.256 nvme0n3: ios=445/0, merge=0/0, ticks=2539/0, in_queue=2539, util=96.07% 00:33:54.256 nvme0n4: ios=98/0, merge=0/0, ticks=3269/0, in_queue=3269, util=100.00% 00:33:54.517 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:33:54.517 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:33:54.777 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:33:54.777 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:33:54.778 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:33:54.778 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:33:55.037 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:33:55.037 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # wait 319783 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:33:55.298 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1219 -- # local i=0 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1231 -- # return 0 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:33:55.298 nvmf hotplug test: fio failed as expected 00:33:55.298 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:33:55.559 rmmod nvme_tcp 00:33:55.559 rmmod nvme_fabrics 00:33:55.559 rmmod nvme_keyring 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@515 -- # '[' -n 316538 ']' 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@516 -- # killprocess 316538 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@950 -- # '[' -z 316538 ']' 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@954 -- # kill -0 316538 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@955 -- # uname 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 316538 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 316538' 00:33:55.559 killing process with pid 316538 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@969 -- # kill 316538 00:33:55.559 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@974 -- # wait 316538 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-save 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-restore 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:55.820 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:57.766 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:57.766 00:33:57.766 real 0m28.321s 00:33:57.766 user 2m20.305s 00:33:57.766 sys 0m12.206s 00:33:57.766 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:57.766 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:33:57.766 ************************************ 00:33:57.766 END TEST nvmf_fio_target 00:33:57.766 ************************************ 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:58.027 ************************************ 00:33:58.027 START TEST nvmf_bdevio 00:33:58.027 ************************************ 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:33:58.027 * Looking for test storage... 00:33:58.027 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1691 -- # lcov --version 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:58.027 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:33:58.289 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:58.289 --rc genhtml_branch_coverage=1 00:33:58.289 --rc genhtml_function_coverage=1 00:33:58.289 --rc genhtml_legend=1 00:33:58.289 --rc geninfo_all_blocks=1 00:33:58.289 --rc geninfo_unexecuted_blocks=1 00:33:58.289 00:33:58.289 ' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:33:58.289 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:58.289 --rc genhtml_branch_coverage=1 00:33:58.289 --rc genhtml_function_coverage=1 00:33:58.289 --rc genhtml_legend=1 00:33:58.289 --rc geninfo_all_blocks=1 00:33:58.289 --rc geninfo_unexecuted_blocks=1 00:33:58.289 00:33:58.289 ' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:33:58.289 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:58.289 --rc genhtml_branch_coverage=1 00:33:58.289 --rc genhtml_function_coverage=1 00:33:58.289 --rc genhtml_legend=1 00:33:58.289 --rc geninfo_all_blocks=1 00:33:58.289 --rc geninfo_unexecuted_blocks=1 00:33:58.289 00:33:58.289 ' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:33:58.289 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:58.289 --rc genhtml_branch_coverage=1 00:33:58.289 --rc genhtml_function_coverage=1 00:33:58.289 --rc genhtml_legend=1 00:33:58.289 --rc geninfo_all_blocks=1 00:33:58.289 --rc geninfo_unexecuted_blocks=1 00:33:58.289 00:33:58.289 ' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@474 -- # prepare_net_devs 00:33:58.289 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@436 -- # local -g is_hw=no 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@438 -- # remove_spdk_ns 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:33:58.290 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:06.435 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:34:06.436 Found 0000:31:00.0 (0x8086 - 0x159b) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:34:06.436 Found 0000:31:00.1 (0x8086 - 0x159b) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:34:06.436 Found net devices under 0000:31:00.0: cvl_0_0 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:34:06.436 Found net devices under 0000:31:00.1: cvl_0_1 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # is_hw=yes 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:06.436 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:06.436 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.642 ms 00:34:06.436 00:34:06.436 --- 10.0.0.2 ping statistics --- 00:34:06.436 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:06.436 rtt min/avg/max/mdev = 0.642/0.642/0.642/0.000 ms 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:06.436 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:06.436 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.313 ms 00:34:06.436 00:34:06.436 --- 10.0.0.1 ping statistics --- 00:34:06.436 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:06.436 rtt min/avg/max/mdev = 0.313/0.313/0.313/0.000 ms 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@448 -- # return 0 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:34:06.436 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@507 -- # nvmfpid=325221 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@508 -- # waitforlisten 325221 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x78 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@831 -- # '[' -z 325221 ']' 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:06.437 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:06.437 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.437 [2024-10-17 17:55:13.655128] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:34:06.437 [2024-10-17 17:55:13.656272] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:34:06.437 [2024-10-17 17:55:13.656319] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:06.437 [2024-10-17 17:55:13.747712] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:34:06.437 [2024-10-17 17:55:13.798187] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:06.437 [2024-10-17 17:55:13.798233] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:06.437 [2024-10-17 17:55:13.798246] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:06.437 [2024-10-17 17:55:13.798253] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:06.437 [2024-10-17 17:55:13.798259] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:06.437 [2024-10-17 17:55:13.800292] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:34:06.437 [2024-10-17 17:55:13.800440] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:34:06.437 [2024-10-17 17:55:13.800599] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:34:06.437 [2024-10-17 17:55:13.800599] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:34:06.437 [2024-10-17 17:55:13.884326] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:34:06.437 [2024-10-17 17:55:13.885114] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:34:06.437 [2024-10-17 17:55:13.885732] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:34:06.437 [2024-10-17 17:55:13.886015] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:34:06.437 [2024-10-17 17:55:13.886053] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@864 -- # return 0 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.739 [2024-10-17 17:55:14.521467] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.739 Malloc0 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:06.739 [2024-10-17 17:55:14.613805] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@558 -- # config=() 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@558 -- # local subsystem config 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:34:06.739 { 00:34:06.739 "params": { 00:34:06.739 "name": "Nvme$subsystem", 00:34:06.739 "trtype": "$TEST_TRANSPORT", 00:34:06.739 "traddr": "$NVMF_FIRST_TARGET_IP", 00:34:06.739 "adrfam": "ipv4", 00:34:06.739 "trsvcid": "$NVMF_PORT", 00:34:06.739 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:34:06.739 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:34:06.739 "hdgst": ${hdgst:-false}, 00:34:06.739 "ddgst": ${ddgst:-false} 00:34:06.739 }, 00:34:06.739 "method": "bdev_nvme_attach_controller" 00:34:06.739 } 00:34:06.739 EOF 00:34:06.739 )") 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@580 -- # cat 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@582 -- # jq . 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@583 -- # IFS=, 00:34:06.739 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:34:06.739 "params": { 00:34:06.739 "name": "Nvme1", 00:34:06.739 "trtype": "tcp", 00:34:06.739 "traddr": "10.0.0.2", 00:34:06.739 "adrfam": "ipv4", 00:34:06.739 "trsvcid": "4420", 00:34:06.739 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:34:06.739 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:34:06.739 "hdgst": false, 00:34:06.739 "ddgst": false 00:34:06.739 }, 00:34:06.739 "method": "bdev_nvme_attach_controller" 00:34:06.739 }' 00:34:07.058 [2024-10-17 17:55:14.671266] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:34:07.058 [2024-10-17 17:55:14.671328] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid325307 ] 00:34:07.058 [2024-10-17 17:55:14.755171] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:34:07.058 [2024-10-17 17:55:14.812469] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:07.058 [2024-10-17 17:55:14.812609] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:07.058 [2024-10-17 17:55:14.812609] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:34:07.348 I/O targets: 00:34:07.348 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:34:07.348 00:34:07.348 00:34:07.348 CUnit - A unit testing framework for C - Version 2.1-3 00:34:07.348 http://cunit.sourceforge.net/ 00:34:07.348 00:34:07.348 00:34:07.348 Suite: bdevio tests on: Nvme1n1 00:34:07.348 Test: blockdev write read block ...passed 00:34:07.348 Test: blockdev write zeroes read block ...passed 00:34:07.348 Test: blockdev write zeroes read no split ...passed 00:34:07.348 Test: blockdev write zeroes read split ...passed 00:34:07.348 Test: blockdev write zeroes read split partial ...passed 00:34:07.348 Test: blockdev reset ...[2024-10-17 17:55:15.094208] nvme_ctrlr.c:1724:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:34:07.348 [2024-10-17 17:55:15.094315] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1b47eb0 (9): Bad file descriptor 00:34:07.348 [2024-10-17 17:55:15.146571] bdev_nvme.c:2184:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:34:07.348 passed 00:34:07.348 Test: blockdev write read 8 blocks ...passed 00:34:07.348 Test: blockdev write read size > 128k ...passed 00:34:07.348 Test: blockdev write read invalid size ...passed 00:34:07.348 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:34:07.348 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:34:07.348 Test: blockdev write read max offset ...passed 00:34:07.609 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:34:07.609 Test: blockdev writev readv 8 blocks ...passed 00:34:07.609 Test: blockdev writev readv 30 x 1block ...passed 00:34:07.609 Test: blockdev writev readv block ...passed 00:34:07.609 Test: blockdev writev readv size > 128k ...passed 00:34:07.609 Test: blockdev writev readv size > 128k in two iovs ...passed 00:34:07.609 Test: blockdev comparev and writev ...[2024-10-17 17:55:15.414484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.609 [2024-10-17 17:55:15.414534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.414552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.610 [2024-10-17 17:55:15.414561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.415177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.610 [2024-10-17 17:55:15.415190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.415204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.610 [2024-10-17 17:55:15.415212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.415865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.610 [2024-10-17 17:55:15.415876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.415890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.610 [2024-10-17 17:55:15.415898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.416541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.610 [2024-10-17 17:55:15.416552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.416566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:34:07.610 [2024-10-17 17:55:15.416574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:34:07.610 passed 00:34:07.610 Test: blockdev nvme passthru rw ...passed 00:34:07.610 Test: blockdev nvme passthru vendor specific ...[2024-10-17 17:55:15.500410] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:34:07.610 [2024-10-17 17:55:15.500427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.500824] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:34:07.610 [2024-10-17 17:55:15.500843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.501221] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:34:07.610 [2024-10-17 17:55:15.501231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:34:07.610 [2024-10-17 17:55:15.501639] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:34:07.610 [2024-10-17 17:55:15.501651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:34:07.610 passed 00:34:07.610 Test: blockdev nvme admin passthru ...passed 00:34:07.871 Test: blockdev copy ...passed 00:34:07.871 00:34:07.871 Run Summary: Type Total Ran Passed Failed Inactive 00:34:07.871 suites 1 1 n/a 0 0 00:34:07.871 tests 23 23 23 0 0 00:34:07.871 asserts 152 152 152 0 n/a 00:34:07.871 00:34:07.871 Elapsed time = 1.176 seconds 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@514 -- # nvmfcleanup 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:07.871 rmmod nvme_tcp 00:34:07.871 rmmod nvme_fabrics 00:34:07.871 rmmod nvme_keyring 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:34:07.871 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@515 -- # '[' -n 325221 ']' 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@516 -- # killprocess 325221 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@950 -- # '[' -z 325221 ']' 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@954 -- # kill -0 325221 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@955 -- # uname 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 325221 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@956 -- # process_name=reactor_3 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@960 -- # '[' reactor_3 = sudo ']' 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@968 -- # echo 'killing process with pid 325221' 00:34:08.132 killing process with pid 325221 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@969 -- # kill 325221 00:34:08.132 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@974 -- # wait 325221 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-save 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-restore 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:08.392 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:10.304 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:10.304 00:34:10.304 real 0m12.397s 00:34:10.304 user 0m9.408s 00:34:10.304 sys 0m6.631s 00:34:10.304 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:10.304 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:34:10.304 ************************************ 00:34:10.304 END TEST nvmf_bdevio 00:34:10.304 ************************************ 00:34:10.304 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:34:10.304 00:34:10.304 real 5m2.134s 00:34:10.304 user 10m18.161s 00:34:10.304 sys 2m8.614s 00:34:10.304 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:10.304 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:34:10.304 ************************************ 00:34:10.304 END TEST nvmf_target_core_interrupt_mode 00:34:10.304 ************************************ 00:34:10.565 17:55:18 nvmf_tcp -- nvmf/nvmf.sh@21 -- # run_test nvmf_interrupt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:34:10.565 17:55:18 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:34:10.565 17:55:18 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:10.565 17:55:18 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:10.565 ************************************ 00:34:10.565 START TEST nvmf_interrupt 00:34:10.565 ************************************ 00:34:10.565 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:34:10.565 * Looking for test storage... 00:34:10.566 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1691 -- # lcov --version 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # IFS=.-: 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # read -ra ver1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # IFS=.-: 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # read -ra ver2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@338 -- # local 'op=<' 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@340 -- # ver1_l=2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@341 -- # ver2_l=1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@344 -- # case "$op" in 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@345 -- # : 1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # decimal 1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # ver1[v]=1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # decimal 2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # ver2[v]=2 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # return 0 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:34:10.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.566 --rc genhtml_branch_coverage=1 00:34:10.566 --rc genhtml_function_coverage=1 00:34:10.566 --rc genhtml_legend=1 00:34:10.566 --rc geninfo_all_blocks=1 00:34:10.566 --rc geninfo_unexecuted_blocks=1 00:34:10.566 00:34:10.566 ' 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:34:10.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.566 --rc genhtml_branch_coverage=1 00:34:10.566 --rc genhtml_function_coverage=1 00:34:10.566 --rc genhtml_legend=1 00:34:10.566 --rc geninfo_all_blocks=1 00:34:10.566 --rc geninfo_unexecuted_blocks=1 00:34:10.566 00:34:10.566 ' 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:34:10.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.566 --rc genhtml_branch_coverage=1 00:34:10.566 --rc genhtml_function_coverage=1 00:34:10.566 --rc genhtml_legend=1 00:34:10.566 --rc geninfo_all_blocks=1 00:34:10.566 --rc geninfo_unexecuted_blocks=1 00:34:10.566 00:34:10.566 ' 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:34:10.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.566 --rc genhtml_branch_coverage=1 00:34:10.566 --rc genhtml_function_coverage=1 00:34:10.566 --rc genhtml_legend=1 00:34:10.566 --rc geninfo_all_blocks=1 00:34:10.566 --rc geninfo_unexecuted_blocks=1 00:34:10.566 00:34:10.566 ' 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # uname -s 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:10.566 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@15 -- # shopt -s extglob 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@5 -- # export PATH 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@51 -- # : 0 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/interrupt/common.sh 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@12 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@14 -- # nvmftestinit 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@474 -- # prepare_net_devs 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@436 -- # local -g is_hw=no 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@438 -- # remove_spdk_ns 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:34:10.827 17:55:18 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@309 -- # xtrace_disable 00:34:10.828 17:55:18 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # pci_devs=() 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # net_devs=() 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # e810=() 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # local -ga e810 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # x722=() 00:34:18.972 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # local -ga x722 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # mlx=() 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # local -ga mlx 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:34:18.973 Found 0000:31:00.0 (0x8086 - 0x159b) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:34:18.973 Found 0000:31:00.1 (0x8086 - 0x159b) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:34:18.973 Found net devices under 0000:31:00.0: cvl_0_0 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:34:18.973 Found net devices under 0000:31:00.1: cvl_0_1 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # is_hw=yes 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:18.973 17:55:25 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:18.973 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:18.973 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.645 ms 00:34:18.973 00:34:18.973 --- 10.0.0.2 ping statistics --- 00:34:18.973 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:18.973 rtt min/avg/max/mdev = 0.645/0.645/0.645/0.000 ms 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:18.973 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:18.973 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.286 ms 00:34:18.973 00:34:18.973 --- 10.0.0.1 ping statistics --- 00:34:18.973 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:18.973 rtt min/avg/max/mdev = 0.286/0.286/0.286/0.000 ms 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@448 -- # return 0 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@15 -- # nvmfappstart -m 0x3 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@507 -- # nvmfpid=329782 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@508 -- # waitforlisten 329782 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@831 -- # '[' -z 329782 ']' 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:18.973 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:18.973 17:55:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:18.973 [2024-10-17 17:55:26.204046] thread.c:2992:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:34:18.973 [2024-10-17 17:55:26.205594] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:34:18.973 [2024-10-17 17:55:26.205662] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:18.974 [2024-10-17 17:55:26.298075] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:34:18.974 [2024-10-17 17:55:26.349731] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:18.974 [2024-10-17 17:55:26.349781] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:18.974 [2024-10-17 17:55:26.349790] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:18.974 [2024-10-17 17:55:26.349797] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:18.974 [2024-10-17 17:55:26.349803] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:18.974 [2024-10-17 17:55:26.351663] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:18.974 [2024-10-17 17:55:26.351666] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:18.974 [2024-10-17 17:55:26.429127] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:34:18.974 [2024-10-17 17:55:26.429715] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:34:18.974 [2024-10-17 17:55:26.430010] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@864 -- # return 0 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@16 -- # setup_bdev_aio 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # uname -s 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # [[ Linux != \F\r\e\e\B\S\D ]] 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@78 -- # dd if=/dev/zero of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile bs=2048 count=5000 00:34:19.235 5000+0 records in 00:34:19.235 5000+0 records out 00:34:19.235 10240000 bytes (10 MB, 9.8 MiB) copied, 0.0182015 s, 563 MB/s 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@79 -- # rpc_cmd bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile AIO0 2048 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:19.235 AIO0 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -q 256 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:19.235 [2024-10-17 17:55:27.136888] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:19.235 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:19.499 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:19.499 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 AIO0 00:34:19.499 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:19.500 [2024-10-17 17:55:27.181506] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 329782 0 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 329782 0 idle 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329782 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.31 reactor_0' 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329782 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.31 reactor_0 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 329782 1 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 329782 1 idle 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:19.500 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329833 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.00 reactor_1' 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329833 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.00 reactor_1 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:19.763 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@28 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@35 -- # perf_pid=330069 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 256 -o 4096 -w randrw -M 30 -t 10 -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 329782 0 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 329782 0 busy 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:19.764 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329782 root 20 0 128.2g 43776 32256 R 99.9 0.0 0:00.50 reactor_0' 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329782 root 20 0 128.2g 43776 32256 R 99.9 0.0 0:00.50 reactor_0 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=99.9 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=99 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 329782 1 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 329782 1 busy 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329833 root 20 0 128.2g 43776 32256 R 99.9 0.0 0:00.28 reactor_1' 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329833 root 20 0 128.2g 43776 32256 R 99.9 0.0 0:00.28 reactor_1 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:20.025 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:20.286 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=99.9 00:34:20.286 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=99 00:34:20.286 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:34:20.286 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:34:20.286 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:34:20.286 17:55:27 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:20.286 17:55:27 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@42 -- # wait 330069 00:34:30.287 Initializing NVMe Controllers 00:34:30.287 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:34:30.287 Controller IO queue size 256, less than required. 00:34:30.287 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:34:30.287 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:34:30.287 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:34:30.287 Initialization complete. Launching workers. 00:34:30.287 ======================================================== 00:34:30.287 Latency(us) 00:34:30.287 Device Information : IOPS MiB/s Average min max 00:34:30.287 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 18325.80 71.59 13974.46 3992.85 34068.63 00:34:30.287 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 20330.60 79.42 12593.34 8023.47 28417.26 00:34:30.287 ======================================================== 00:34:30.287 Total : 38656.40 151.00 13248.09 3992.85 34068.63 00:34:30.287 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 329782 0 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 329782 0 idle 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329782 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:20.30 reactor_0' 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329782 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:20.30 reactor_0 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:34:30.287 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 329782 1 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 329782 1 idle 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:30.288 17:55:37 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329833 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:10.01 reactor_1' 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329833 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:10.01 reactor_1 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:30.288 17:55:38 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@50 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:34:30.859 17:55:38 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@51 -- # waitforserial SPDKISFASTANDAWESOME 00:34:30.860 17:55:38 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1198 -- # local i=0 00:34:30.860 17:55:38 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:34:30.860 17:55:38 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:34:30.860 17:55:38 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1205 -- # sleep 2 00:34:33.403 17:55:40 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:34:33.403 17:55:40 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1208 -- # return 0 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 329782 0 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 329782 0 idle 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329782 root 20 0 128.2g 78336 32256 S 6.2 0.1 0:20.69 reactor_0' 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329782 root 20 0 128.2g 78336 32256 S 6.2 0.1 0:20.69 reactor_0 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=6.2 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=6 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 329782 1 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 329782 1 idle 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=329782 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 329782 -w 256 00:34:33.404 17:55:40 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 329833 root 20 0 128.2g 78336 32256 S 0.0 0.1 0:10.15 reactor_1' 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 329833 root 20 0 128.2g 78336 32256 S 0.0 0.1 0:10.15 reactor_1 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@55 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:34:33.404 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@56 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1219 -- # local i=0 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1231 -- # return 0 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@58 -- # trap - SIGINT SIGTERM EXIT 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@59 -- # nvmftestfini 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@514 -- # nvmfcleanup 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@121 -- # sync 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@124 -- # set +e 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:33.404 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:33.404 rmmod nvme_tcp 00:34:33.664 rmmod nvme_fabrics 00:34:33.664 rmmod nvme_keyring 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@128 -- # set -e 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@129 -- # return 0 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@515 -- # '[' -n 329782 ']' 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@516 -- # killprocess 329782 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@950 -- # '[' -z 329782 ']' 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@954 -- # kill -0 329782 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@955 -- # uname 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 329782 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@968 -- # echo 'killing process with pid 329782' 00:34:33.664 killing process with pid 329782 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@969 -- # kill 329782 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@974 -- # wait 329782 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@297 -- # iptr 00:34:33.664 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@789 -- # iptables-save 00:34:33.665 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:34:33.665 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@789 -- # iptables-restore 00:34:33.665 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:33.665 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:33.665 17:55:41 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:33.665 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:34:33.665 17:55:41 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:36.210 17:55:43 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:36.210 00:34:36.210 real 0m25.390s 00:34:36.210 user 0m40.191s 00:34:36.210 sys 0m9.916s 00:34:36.210 17:55:43 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:36.210 17:55:43 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:34:36.210 ************************************ 00:34:36.210 END TEST nvmf_interrupt 00:34:36.210 ************************************ 00:34:36.210 00:34:36.210 real 29m59.364s 00:34:36.210 user 61m7.169s 00:34:36.210 sys 10m18.793s 00:34:36.210 17:55:43 nvmf_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:36.210 17:55:43 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:36.210 ************************************ 00:34:36.210 END TEST nvmf_tcp 00:34:36.210 ************************************ 00:34:36.210 17:55:43 -- spdk/autotest.sh@281 -- # [[ 0 -eq 0 ]] 00:34:36.210 17:55:43 -- spdk/autotest.sh@282 -- # run_test spdkcli_nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:34:36.210 17:55:43 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:34:36.210 17:55:43 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:36.210 17:55:43 -- common/autotest_common.sh@10 -- # set +x 00:34:36.210 ************************************ 00:34:36.210 START TEST spdkcli_nvmf_tcp 00:34:36.210 ************************************ 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:34:36.210 * Looking for test storage... 00:34:36.210 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1691 -- # lcov --version 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:34:36.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:36.210 --rc genhtml_branch_coverage=1 00:34:36.210 --rc genhtml_function_coverage=1 00:34:36.210 --rc genhtml_legend=1 00:34:36.210 --rc geninfo_all_blocks=1 00:34:36.210 --rc geninfo_unexecuted_blocks=1 00:34:36.210 00:34:36.210 ' 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:34:36.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:36.210 --rc genhtml_branch_coverage=1 00:34:36.210 --rc genhtml_function_coverage=1 00:34:36.210 --rc genhtml_legend=1 00:34:36.210 --rc geninfo_all_blocks=1 00:34:36.210 --rc geninfo_unexecuted_blocks=1 00:34:36.210 00:34:36.210 ' 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:34:36.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:36.210 --rc genhtml_branch_coverage=1 00:34:36.210 --rc genhtml_function_coverage=1 00:34:36.210 --rc genhtml_legend=1 00:34:36.210 --rc geninfo_all_blocks=1 00:34:36.210 --rc geninfo_unexecuted_blocks=1 00:34:36.210 00:34:36.210 ' 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:34:36.210 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:36.210 --rc genhtml_branch_coverage=1 00:34:36.210 --rc genhtml_function_coverage=1 00:34:36.210 --rc genhtml_legend=1 00:34:36.210 --rc geninfo_all_blocks=1 00:34:36.210 --rc geninfo_unexecuted_blocks=1 00:34:36.210 00:34:36.210 ' 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:36.210 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # uname -s 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@15 -- # shopt -s extglob 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- paths/export.sh@5 -- # export PATH 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@51 -- # : 0 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:36.211 17:55:43 spdkcli_nvmf_tcp -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:34:36.211 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@12 -- # MATCH_FILE=spdkcli_nvmf.test 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@13 -- # SPDKCLI_BRANCH=/nvmf 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@15 -- # trap cleanup EXIT 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@17 -- # timing_enter run_nvmf_tgt 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@18 -- # run_nvmf_tgt 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@33 -- # nvmf_tgt_pid=333262 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@34 -- # waitforlisten 333262 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@831 -- # '[' -z 333262 ']' 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- spdkcli/common.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x3 -p 0 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:36.211 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:36.211 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:36.211 [2024-10-17 17:55:44.081396] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:34:36.211 [2024-10-17 17:55:44.081463] [ DPDK EAL parameters: nvmf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid333262 ] 00:34:36.472 [2024-10-17 17:55:44.166740] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:34:36.472 [2024-10-17 17:55:44.222379] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:36.472 [2024-10-17 17:55:44.222384] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@864 -- # return 0 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@19 -- # timing_exit run_nvmf_tgt 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@21 -- # NVMF_TARGET_IP=127.0.0.1 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@22 -- # [[ tcp == \r\d\m\a ]] 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@27 -- # timing_enter spdkcli_create_nvmf_config 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:37.043 17:55:44 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 32 512 Malloc1'\'' '\''Malloc1'\'' True 00:34:37.043 '\''/bdevs/malloc create 32 512 Malloc2'\'' '\''Malloc2'\'' True 00:34:37.043 '\''/bdevs/malloc create 32 512 Malloc3'\'' '\''Malloc3'\'' True 00:34:37.043 '\''/bdevs/malloc create 32 512 Malloc4'\'' '\''Malloc4'\'' True 00:34:37.043 '\''/bdevs/malloc create 32 512 Malloc5'\'' '\''Malloc5'\'' True 00:34:37.043 '\''/bdevs/malloc create 32 512 Malloc6'\'' '\''Malloc6'\'' True 00:34:37.043 '\''nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192'\'' '\'''\'' True 00:34:37.043 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1'\'' '\''Malloc3'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2'\'' '\''Malloc4'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:34:37.043 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2'\'' '\''Malloc2'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:34:37.043 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1'\'' '\''Malloc1'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True'\'' '\''Allow any host'\'' 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False'\'' '\''Allow any host'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4'\'' '\''127.0.0.1:4262'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5'\'' '\''Malloc5'\'' True 00:34:37.043 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6'\'' '\''Malloc6'\'' True 00:34:37.043 '\''/nvmf/referral create tcp 127.0.0.2 4030 IPv4'\'' 00:34:37.043 ' 00:34:40.338 [2024-10-17 17:55:47.602770] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:41.278 [2024-10-17 17:55:48.958909] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4260 *** 00:34:43.822 [2024-10-17 17:55:51.482262] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4261 *** 00:34:46.369 [2024-10-17 17:55:53.700557] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4262 *** 00:34:47.752 Executing command: ['/bdevs/malloc create 32 512 Malloc1', 'Malloc1', True] 00:34:47.752 Executing command: ['/bdevs/malloc create 32 512 Malloc2', 'Malloc2', True] 00:34:47.752 Executing command: ['/bdevs/malloc create 32 512 Malloc3', 'Malloc3', True] 00:34:47.752 Executing command: ['/bdevs/malloc create 32 512 Malloc4', 'Malloc4', True] 00:34:47.752 Executing command: ['/bdevs/malloc create 32 512 Malloc5', 'Malloc5', True] 00:34:47.752 Executing command: ['/bdevs/malloc create 32 512 Malloc6', 'Malloc6', True] 00:34:47.752 Executing command: ['nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192', '', True] 00:34:47.752 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode1', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1', 'Malloc3', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2', 'Malloc4', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:34:47.752 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2', 'Malloc2', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:34:47.752 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1', 'Malloc1', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1', 'nqn.2014-08.org.spdk:cnode1', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True', 'Allow any host', False] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False', 'Allow any host', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4', '127.0.0.1:4262', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5', 'Malloc5', True] 00:34:47.752 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6', 'Malloc6', True] 00:34:47.752 Executing command: ['/nvmf/referral create tcp 127.0.0.2 4030 IPv4', False] 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@66 -- # timing_exit spdkcli_create_nvmf_config 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@68 -- # timing_enter spdkcli_check_match 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@69 -- # check_match 00:34:47.752 17:55:55 spdkcli_nvmf_tcp -- spdkcli/common.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdkcli.py ll /nvmf 00:34:48.013 17:55:55 spdkcli_nvmf_tcp -- spdkcli/common.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/match/match /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test.match 00:34:48.274 17:55:55 spdkcli_nvmf_tcp -- spdkcli/common.sh@46 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test 00:34:48.274 17:55:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@70 -- # timing_exit spdkcli_check_match 00:34:48.274 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:48.274 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:48.274 17:55:55 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@72 -- # timing_enter spdkcli_clear_nvmf_config 00:34:48.274 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:48.274 17:55:55 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:48.274 17:55:56 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1'\'' '\''Malloc3'\'' 00:34:48.274 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all'\'' '\''Malloc4'\'' 00:34:48.274 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:34:48.274 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' 00:34:48.274 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262'\'' '\''127.0.0.1:4262'\'' 00:34:48.274 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all'\'' '\''127.0.0.1:4261'\'' 00:34:48.274 '\''/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3'\'' '\''nqn.2014-08.org.spdk:cnode3'\'' 00:34:48.274 '\''/nvmf/subsystem delete_all'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:34:48.274 '\''/bdevs/malloc delete Malloc6'\'' '\''Malloc6'\'' 00:34:48.274 '\''/bdevs/malloc delete Malloc5'\'' '\''Malloc5'\'' 00:34:48.274 '\''/bdevs/malloc delete Malloc4'\'' '\''Malloc4'\'' 00:34:48.274 '\''/bdevs/malloc delete Malloc3'\'' '\''Malloc3'\'' 00:34:48.274 '\''/bdevs/malloc delete Malloc2'\'' '\''Malloc2'\'' 00:34:48.274 '\''/bdevs/malloc delete Malloc1'\'' '\''Malloc1'\'' 00:34:48.274 ' 00:34:54.854 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1', 'Malloc3', False] 00:34:54.854 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all', 'Malloc4', False] 00:34:54.854 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', False] 00:34:54.854 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all', 'nqn.2014-08.org.spdk:cnode1', False] 00:34:54.854 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262', '127.0.0.1:4262', False] 00:34:54.854 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all', '127.0.0.1:4261', False] 00:34:54.854 Executing command: ['/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3', 'nqn.2014-08.org.spdk:cnode3', False] 00:34:54.854 Executing command: ['/nvmf/subsystem delete_all', 'nqn.2014-08.org.spdk:cnode2', False] 00:34:54.854 Executing command: ['/bdevs/malloc delete Malloc6', 'Malloc6', False] 00:34:54.854 Executing command: ['/bdevs/malloc delete Malloc5', 'Malloc5', False] 00:34:54.854 Executing command: ['/bdevs/malloc delete Malloc4', 'Malloc4', False] 00:34:54.854 Executing command: ['/bdevs/malloc delete Malloc3', 'Malloc3', False] 00:34:54.854 Executing command: ['/bdevs/malloc delete Malloc2', 'Malloc2', False] 00:34:54.854 Executing command: ['/bdevs/malloc delete Malloc1', 'Malloc1', False] 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@88 -- # timing_exit spdkcli_clear_nvmf_config 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@90 -- # killprocess 333262 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@950 -- # '[' -z 333262 ']' 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # kill -0 333262 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@955 -- # uname 00:34:54.854 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 333262 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 333262' 00:34:54.855 killing process with pid 333262 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@969 -- # kill 333262 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@974 -- # wait 333262 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@1 -- # cleanup 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- spdkcli/common.sh@10 -- # '[' -n '' ']' 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- spdkcli/common.sh@13 -- # '[' -n 333262 ']' 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- spdkcli/common.sh@14 -- # killprocess 333262 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@950 -- # '[' -z 333262 ']' 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # kill -0 333262 00:34:54.855 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (333262) - No such process 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@977 -- # echo 'Process with pid 333262 is not found' 00:34:54.855 Process with pid 333262 is not found 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- spdkcli/common.sh@22 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_nvmf.test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:34:54.855 00:34:54.855 real 0m18.087s 00:34:54.855 user 0m40.132s 00:34:54.855 sys 0m0.903s 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:54.855 17:56:01 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:34:54.855 ************************************ 00:34:54.855 END TEST spdkcli_nvmf_tcp 00:34:54.855 ************************************ 00:34:54.855 17:56:01 -- spdk/autotest.sh@283 -- # run_test nvmf_identify_passthru /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:34:54.855 17:56:01 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:34:54.855 17:56:01 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:54.855 17:56:01 -- common/autotest_common.sh@10 -- # set +x 00:34:54.855 ************************************ 00:34:54.855 START TEST nvmf_identify_passthru 00:34:54.855 ************************************ 00:34:54.855 17:56:01 nvmf_identify_passthru -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:34:54.855 * Looking for test storage... 00:34:54.855 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1691 -- # lcov --version 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@336 -- # IFS=.-: 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@336 -- # read -ra ver1 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@337 -- # IFS=.-: 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@337 -- # read -ra ver2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@338 -- # local 'op=<' 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@340 -- # ver1_l=2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@341 -- # ver2_l=1 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@344 -- # case "$op" in 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@345 -- # : 1 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@365 -- # decimal 1 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=1 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 1 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@365 -- # ver1[v]=1 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@366 -- # decimal 2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@366 -- # ver2[v]=2 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@368 -- # return 0 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:34:54.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:54.855 --rc genhtml_branch_coverage=1 00:34:54.855 --rc genhtml_function_coverage=1 00:34:54.855 --rc genhtml_legend=1 00:34:54.855 --rc geninfo_all_blocks=1 00:34:54.855 --rc geninfo_unexecuted_blocks=1 00:34:54.855 00:34:54.855 ' 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:34:54.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:54.855 --rc genhtml_branch_coverage=1 00:34:54.855 --rc genhtml_function_coverage=1 00:34:54.855 --rc genhtml_legend=1 00:34:54.855 --rc geninfo_all_blocks=1 00:34:54.855 --rc geninfo_unexecuted_blocks=1 00:34:54.855 00:34:54.855 ' 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:34:54.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:54.855 --rc genhtml_branch_coverage=1 00:34:54.855 --rc genhtml_function_coverage=1 00:34:54.855 --rc genhtml_legend=1 00:34:54.855 --rc geninfo_all_blocks=1 00:34:54.855 --rc geninfo_unexecuted_blocks=1 00:34:54.855 00:34:54.855 ' 00:34:54.855 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:34:54.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:54.855 --rc genhtml_branch_coverage=1 00:34:54.855 --rc genhtml_function_coverage=1 00:34:54.855 --rc genhtml_legend=1 00:34:54.855 --rc geninfo_all_blocks=1 00:34:54.855 --rc geninfo_unexecuted_blocks=1 00:34:54.855 00:34:54.855 ' 00:34:54.855 17:56:02 nvmf_identify_passthru -- target/identify_passthru.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@7 -- # uname -s 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:54.855 17:56:02 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:54.855 17:56:02 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.855 17:56:02 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.855 17:56:02 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.855 17:56:02 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:34:54.855 17:56:02 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@51 -- # : 0 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:54.855 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:34:54.856 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:54.856 17:56:02 nvmf_identify_passthru -- target/identify_passthru.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:54.856 17:56:02 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:34:54.856 17:56:02 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:54.856 17:56:02 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:54.856 17:56:02 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:54.856 17:56:02 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.856 17:56:02 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.856 17:56:02 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.856 17:56:02 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:34:54.856 17:56:02 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:54.856 17:56:02 nvmf_identify_passthru -- target/identify_passthru.sh@12 -- # nvmftestinit 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@474 -- # prepare_net_devs 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@436 -- # local -g is_hw=no 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@438 -- # remove_spdk_ns 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:54.856 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:34:54.856 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:34:54.856 17:56:02 nvmf_identify_passthru -- nvmf/common.sh@309 -- # xtrace_disable 00:34:54.856 17:56:02 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@315 -- # pci_devs=() 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@319 -- # net_devs=() 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@320 -- # e810=() 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@320 -- # local -ga e810 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@321 -- # x722=() 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@321 -- # local -ga x722 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@322 -- # mlx=() 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@322 -- # local -ga mlx 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:35:02.994 Found 0000:31:00.0 (0x8086 - 0x159b) 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:02.994 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:35:02.995 Found 0000:31:00.1 (0x8086 - 0x159b) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:35:02.995 Found net devices under 0000:31:00.0: cvl_0_0 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:35:02.995 Found net devices under 0000:31:00.1: cvl_0_1 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@440 -- # is_hw=yes 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:02.995 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:02.995 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.590 ms 00:35:02.995 00:35:02.995 --- 10.0.0.2 ping statistics --- 00:35:02.995 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:02.995 rtt min/avg/max/mdev = 0.590/0.590/0.590/0.000 ms 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:02.995 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:02.995 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.270 ms 00:35:02.995 00:35:02.995 --- 10.0.0.1 ping statistics --- 00:35:02.995 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:02.995 rtt min/avg/max/mdev = 0.270/0.270/0.270/0.000 ms 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@448 -- # return 0 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:35:02.995 17:56:09 nvmf_identify_passthru -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:35:02.995 17:56:09 nvmf_identify_passthru -- target/identify_passthru.sh@14 -- # timing_enter nvme_identify 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@724 -- # xtrace_disable 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:02.995 17:56:09 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # get_first_nvme_bdf 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1507 -- # bdfs=() 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1507 -- # local bdfs 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1508 -- # bdfs=($(get_nvme_bdfs)) 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1508 -- # get_nvme_bdfs 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1496 -- # bdfs=() 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1496 -- # local bdfs 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1497 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1498 -- # (( 1 == 0 )) 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:65:00.0 00:35:02.995 17:56:09 nvmf_identify_passthru -- common/autotest_common.sh@1510 -- # echo 0000:65:00.0 00:35:02.995 17:56:09 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # bdf=0000:65:00.0 00:35:02.995 17:56:09 nvmf_identify_passthru -- target/identify_passthru.sh@17 -- # '[' -z 0000:65:00.0 ']' 00:35:02.995 17:56:09 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:65:00.0' -i 0 00:35:02.995 17:56:09 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # grep 'Serial Number:' 00:35:02.995 17:56:09 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # awk '{print $3}' 00:35:02.995 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # nvme_serial_number=S64GNE0R605500 00:35:02.995 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:65:00.0' -i 0 00:35:02.995 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # grep 'Model Number:' 00:35:02.995 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # awk '{print $3}' 00:35:02.995 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # nvme_model_number=SAMSUNG 00:35:02.995 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@26 -- # timing_exit nvme_identify 00:35:02.995 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@730 -- # xtrace_disable 00:35:02.995 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:03.257 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@28 -- # timing_enter start_nvmf_tgt 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@724 -- # xtrace_disable 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:03.257 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@31 -- # nvmfpid=340692 00:35:03.257 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:35:03.257 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@30 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:35:03.257 17:56:10 nvmf_identify_passthru -- target/identify_passthru.sh@35 -- # waitforlisten 340692 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@831 -- # '[' -z 340692 ']' 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@836 -- # local max_retries=100 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:03.257 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@840 -- # xtrace_disable 00:35:03.257 17:56:10 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:03.257 [2024-10-17 17:56:11.024394] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:35:03.257 [2024-10-17 17:56:11.024472] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:03.257 [2024-10-17 17:56:11.116506] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:35:03.257 [2024-10-17 17:56:11.170647] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:03.257 [2024-10-17 17:56:11.170713] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:03.257 [2024-10-17 17:56:11.170723] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:03.257 [2024-10-17 17:56:11.170731] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:03.257 [2024-10-17 17:56:11.170737] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:03.257 [2024-10-17 17:56:11.172983] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:03.257 [2024-10-17 17:56:11.173148] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:35:03.257 [2024-10-17 17:56:11.173308] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:35:03.257 [2024-10-17 17:56:11.173309] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@864 -- # return 0 00:35:04.199 17:56:11 nvmf_identify_passthru -- target/identify_passthru.sh@36 -- # rpc_cmd -v nvmf_set_config --passthru-identify-ctrlr 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.199 INFO: Log level set to 20 00:35:04.199 INFO: Requests: 00:35:04.199 { 00:35:04.199 "jsonrpc": "2.0", 00:35:04.199 "method": "nvmf_set_config", 00:35:04.199 "id": 1, 00:35:04.199 "params": { 00:35:04.199 "admin_cmd_passthru": { 00:35:04.199 "identify_ctrlr": true 00:35:04.199 } 00:35:04.199 } 00:35:04.199 } 00:35:04.199 00:35:04.199 INFO: response: 00:35:04.199 { 00:35:04.199 "jsonrpc": "2.0", 00:35:04.199 "id": 1, 00:35:04.199 "result": true 00:35:04.199 } 00:35:04.199 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.199 17:56:11 nvmf_identify_passthru -- target/identify_passthru.sh@37 -- # rpc_cmd -v framework_start_init 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.199 INFO: Setting log level to 20 00:35:04.199 INFO: Setting log level to 20 00:35:04.199 INFO: Log level set to 20 00:35:04.199 INFO: Log level set to 20 00:35:04.199 INFO: Requests: 00:35:04.199 { 00:35:04.199 "jsonrpc": "2.0", 00:35:04.199 "method": "framework_start_init", 00:35:04.199 "id": 1 00:35:04.199 } 00:35:04.199 00:35:04.199 INFO: Requests: 00:35:04.199 { 00:35:04.199 "jsonrpc": "2.0", 00:35:04.199 "method": "framework_start_init", 00:35:04.199 "id": 1 00:35:04.199 } 00:35:04.199 00:35:04.199 [2024-10-17 17:56:11.927975] nvmf_tgt.c: 462:nvmf_tgt_advance_state: *NOTICE*: Custom identify ctrlr handler enabled 00:35:04.199 INFO: response: 00:35:04.199 { 00:35:04.199 "jsonrpc": "2.0", 00:35:04.199 "id": 1, 00:35:04.199 "result": true 00:35:04.199 } 00:35:04.199 00:35:04.199 INFO: response: 00:35:04.199 { 00:35:04.199 "jsonrpc": "2.0", 00:35:04.199 "id": 1, 00:35:04.199 "result": true 00:35:04.199 } 00:35:04.199 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.199 17:56:11 nvmf_identify_passthru -- target/identify_passthru.sh@38 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.199 INFO: Setting log level to 40 00:35:04.199 INFO: Setting log level to 40 00:35:04.199 INFO: Setting log level to 40 00:35:04.199 [2024-10-17 17:56:11.941565] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.199 17:56:11 nvmf_identify_passthru -- target/identify_passthru.sh@39 -- # timing_exit start_nvmf_tgt 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@730 -- # xtrace_disable 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.199 17:56:11 nvmf_identify_passthru -- target/identify_passthru.sh@41 -- # rpc_cmd bdev_nvme_attach_controller -b Nvme0 -t PCIe -a 0000:65:00.0 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.199 17:56:11 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.460 Nvme0n1 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.460 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@42 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 1 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.460 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@43 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.460 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@44 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.460 [2024-10-17 17:56:12.347960] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.460 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@46 -- # rpc_cmd nvmf_get_subsystems 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.460 [ 00:35:04.460 { 00:35:04.460 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:35:04.460 "subtype": "Discovery", 00:35:04.460 "listen_addresses": [], 00:35:04.460 "allow_any_host": true, 00:35:04.460 "hosts": [] 00:35:04.460 }, 00:35:04.460 { 00:35:04.460 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:35:04.460 "subtype": "NVMe", 00:35:04.460 "listen_addresses": [ 00:35:04.460 { 00:35:04.460 "trtype": "TCP", 00:35:04.460 "adrfam": "IPv4", 00:35:04.460 "traddr": "10.0.0.2", 00:35:04.460 "trsvcid": "4420" 00:35:04.460 } 00:35:04.460 ], 00:35:04.460 "allow_any_host": true, 00:35:04.460 "hosts": [], 00:35:04.460 "serial_number": "SPDK00000000000001", 00:35:04.460 "model_number": "SPDK bdev Controller", 00:35:04.460 "max_namespaces": 1, 00:35:04.460 "min_cntlid": 1, 00:35:04.460 "max_cntlid": 65519, 00:35:04.460 "namespaces": [ 00:35:04.460 { 00:35:04.460 "nsid": 1, 00:35:04.460 "bdev_name": "Nvme0n1", 00:35:04.460 "name": "Nvme0n1", 00:35:04.460 "nguid": "36344730526055000025384500000031", 00:35:04.460 "uuid": "36344730-5260-5500-0025-384500000031" 00:35:04.460 } 00:35:04.460 ] 00:35:04.460 } 00:35:04.460 ] 00:35:04.460 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.460 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:35:04.460 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # grep 'Serial Number:' 00:35:04.460 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # awk '{print $3}' 00:35:04.721 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # nvmf_serial_number=S64GNE0R605500 00:35:04.721 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:35:04.721 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # grep 'Model Number:' 00:35:04.721 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # awk '{print $3}' 00:35:04.982 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # nvmf_model_number=SAMSUNG 00:35:04.982 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@63 -- # '[' S64GNE0R605500 '!=' S64GNE0R605500 ']' 00:35:04.982 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@68 -- # '[' SAMSUNG '!=' SAMSUNG ']' 00:35:04.982 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@73 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:35:04.982 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:04.982 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:04.982 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:04.982 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@75 -- # trap - SIGINT SIGTERM EXIT 00:35:04.982 17:56:12 nvmf_identify_passthru -- target/identify_passthru.sh@77 -- # nvmftestfini 00:35:04.982 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@514 -- # nvmfcleanup 00:35:04.982 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@121 -- # sync 00:35:04.982 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:04.982 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@124 -- # set +e 00:35:04.982 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:04.982 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:04.982 rmmod nvme_tcp 00:35:04.982 rmmod nvme_fabrics 00:35:05.243 rmmod nvme_keyring 00:35:05.243 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:05.243 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@128 -- # set -e 00:35:05.243 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@129 -- # return 0 00:35:05.243 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@515 -- # '[' -n 340692 ']' 00:35:05.243 17:56:12 nvmf_identify_passthru -- nvmf/common.sh@516 -- # killprocess 340692 00:35:05.243 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@950 -- # '[' -z 340692 ']' 00:35:05.243 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@954 -- # kill -0 340692 00:35:05.243 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@955 -- # uname 00:35:05.243 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:35:05.243 17:56:12 nvmf_identify_passthru -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 340692 00:35:05.243 17:56:13 nvmf_identify_passthru -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:35:05.243 17:56:13 nvmf_identify_passthru -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:35:05.243 17:56:13 nvmf_identify_passthru -- common/autotest_common.sh@968 -- # echo 'killing process with pid 340692' 00:35:05.243 killing process with pid 340692 00:35:05.243 17:56:13 nvmf_identify_passthru -- common/autotest_common.sh@969 -- # kill 340692 00:35:05.243 17:56:13 nvmf_identify_passthru -- common/autotest_common.sh@974 -- # wait 340692 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@297 -- # iptr 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@789 -- # iptables-save 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@789 -- # iptables-restore 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:05.504 17:56:13 nvmf_identify_passthru -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:05.504 17:56:13 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:35:05.504 17:56:13 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:08.051 17:56:15 nvmf_identify_passthru -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:08.051 00:35:08.051 real 0m13.425s 00:35:08.051 user 0m10.558s 00:35:08.051 sys 0m6.939s 00:35:08.051 17:56:15 nvmf_identify_passthru -- common/autotest_common.sh@1126 -- # xtrace_disable 00:35:08.051 17:56:15 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:35:08.051 ************************************ 00:35:08.051 END TEST nvmf_identify_passthru 00:35:08.051 ************************************ 00:35:08.051 17:56:15 -- spdk/autotest.sh@285 -- # run_test nvmf_dif /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:35:08.051 17:56:15 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:35:08.051 17:56:15 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:08.051 17:56:15 -- common/autotest_common.sh@10 -- # set +x 00:35:08.051 ************************************ 00:35:08.051 START TEST nvmf_dif 00:35:08.051 ************************************ 00:35:08.051 17:56:15 nvmf_dif -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:35:08.051 * Looking for test storage... 00:35:08.051 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:35:08.051 17:56:15 nvmf_dif -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:35:08.051 17:56:15 nvmf_dif -- common/autotest_common.sh@1691 -- # lcov --version 00:35:08.051 17:56:15 nvmf_dif -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:35:08.051 17:56:15 nvmf_dif -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@336 -- # IFS=.-: 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@336 -- # read -ra ver1 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@337 -- # IFS=.-: 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@337 -- # read -ra ver2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@338 -- # local 'op=<' 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@340 -- # ver1_l=2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@341 -- # ver2_l=1 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@344 -- # case "$op" in 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@345 -- # : 1 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@365 -- # decimal 1 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@353 -- # local d=1 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@355 -- # echo 1 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@365 -- # ver1[v]=1 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@366 -- # decimal 2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@353 -- # local d=2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@355 -- # echo 2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@366 -- # ver2[v]=2 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:08.051 17:56:15 nvmf_dif -- scripts/common.sh@368 -- # return 0 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:35:08.052 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.052 --rc genhtml_branch_coverage=1 00:35:08.052 --rc genhtml_function_coverage=1 00:35:08.052 --rc genhtml_legend=1 00:35:08.052 --rc geninfo_all_blocks=1 00:35:08.052 --rc geninfo_unexecuted_blocks=1 00:35:08.052 00:35:08.052 ' 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:35:08.052 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.052 --rc genhtml_branch_coverage=1 00:35:08.052 --rc genhtml_function_coverage=1 00:35:08.052 --rc genhtml_legend=1 00:35:08.052 --rc geninfo_all_blocks=1 00:35:08.052 --rc geninfo_unexecuted_blocks=1 00:35:08.052 00:35:08.052 ' 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:35:08.052 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.052 --rc genhtml_branch_coverage=1 00:35:08.052 --rc genhtml_function_coverage=1 00:35:08.052 --rc genhtml_legend=1 00:35:08.052 --rc geninfo_all_blocks=1 00:35:08.052 --rc geninfo_unexecuted_blocks=1 00:35:08.052 00:35:08.052 ' 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:35:08.052 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:08.052 --rc genhtml_branch_coverage=1 00:35:08.052 --rc genhtml_function_coverage=1 00:35:08.052 --rc genhtml_legend=1 00:35:08.052 --rc geninfo_all_blocks=1 00:35:08.052 --rc geninfo_unexecuted_blocks=1 00:35:08.052 00:35:08.052 ' 00:35:08.052 17:56:15 nvmf_dif -- target/dif.sh@13 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@7 -- # uname -s 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:35:08.052 17:56:15 nvmf_dif -- scripts/common.sh@15 -- # shopt -s extglob 00:35:08.052 17:56:15 nvmf_dif -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:35:08.052 17:56:15 nvmf_dif -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:08.052 17:56:15 nvmf_dif -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:08.052 17:56:15 nvmf_dif -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.052 17:56:15 nvmf_dif -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.052 17:56:15 nvmf_dif -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.052 17:56:15 nvmf_dif -- paths/export.sh@5 -- # export PATH 00:35:08.052 17:56:15 nvmf_dif -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@51 -- # : 0 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:35:08.052 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@55 -- # have_pci_nics=0 00:35:08.052 17:56:15 nvmf_dif -- target/dif.sh@15 -- # NULL_META=16 00:35:08.052 17:56:15 nvmf_dif -- target/dif.sh@15 -- # NULL_BLOCK_SIZE=512 00:35:08.052 17:56:15 nvmf_dif -- target/dif.sh@15 -- # NULL_SIZE=64 00:35:08.052 17:56:15 nvmf_dif -- target/dif.sh@15 -- # NULL_DIF=1 00:35:08.052 17:56:15 nvmf_dif -- target/dif.sh@135 -- # nvmftestinit 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@474 -- # prepare_net_devs 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@436 -- # local -g is_hw=no 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@438 -- # remove_spdk_ns 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:35:08.052 17:56:15 nvmf_dif -- nvmf/common.sh@309 -- # xtrace_disable 00:35:08.052 17:56:15 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@315 -- # pci_devs=() 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@319 -- # net_devs=() 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@320 -- # e810=() 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@320 -- # local -ga e810 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@321 -- # x722=() 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@321 -- # local -ga x722 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@322 -- # mlx=() 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@322 -- # local -ga mlx 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:35:16.189 Found 0000:31:00.0 (0x8086 - 0x159b) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:35:16.189 Found 0000:31:00.1 (0x8086 - 0x159b) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:35:16.189 Found net devices under 0000:31:00.0: cvl_0_0 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:35:16.189 Found net devices under 0000:31:00.1: cvl_0_1 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@440 -- # is_hw=yes 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:16.189 17:56:22 nvmf_dif -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:16.189 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:16.189 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.646 ms 00:35:16.189 00:35:16.189 --- 10.0.0.2 ping statistics --- 00:35:16.189 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:16.189 rtt min/avg/max/mdev = 0.646/0.646/0.646/0.000 ms 00:35:16.189 17:56:23 nvmf_dif -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:16.190 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:16.190 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.296 ms 00:35:16.190 00:35:16.190 --- 10.0.0.1 ping statistics --- 00:35:16.190 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:16.190 rtt min/avg/max/mdev = 0.296/0.296/0.296/0.000 ms 00:35:16.190 17:56:23 nvmf_dif -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:16.190 17:56:23 nvmf_dif -- nvmf/common.sh@448 -- # return 0 00:35:16.190 17:56:23 nvmf_dif -- nvmf/common.sh@476 -- # '[' iso == iso ']' 00:35:16.190 17:56:23 nvmf_dif -- nvmf/common.sh@477 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:35:19.493 0000:80:01.6 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:80:01.7 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:80:01.4 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:80:01.5 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:80:01.2 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:80:01.3 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:80:01.0 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:80:01.1 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:00:01.6 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:65:00.0 (144d a80a): Already using the vfio-pci driver 00:35:19.493 0000:00:01.7 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:00:01.4 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:00:01.5 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:00:01.2 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:00:01.3 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:00:01.0 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 0000:00:01.1 (8086 0b00): Already using the vfio-pci driver 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:35:19.493 17:56:27 nvmf_dif -- target/dif.sh@136 -- # NVMF_TRANSPORT_OPTS+=' --dif-insert-or-strip' 00:35:19.493 17:56:27 nvmf_dif -- target/dif.sh@137 -- # nvmfappstart 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@724 -- # xtrace_disable 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@507 -- # nvmfpid=346911 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@508 -- # waitforlisten 346911 00:35:19.493 17:56:27 nvmf_dif -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@831 -- # '[' -z 346911 ']' 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@836 -- # local max_retries=100 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:19.493 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@840 -- # xtrace_disable 00:35:19.493 17:56:27 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:19.493 [2024-10-17 17:56:27.246972] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:35:19.493 [2024-10-17 17:56:27.247038] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:19.493 [2024-10-17 17:56:27.338522] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:19.493 [2024-10-17 17:56:27.390350] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:19.493 [2024-10-17 17:56:27.390401] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:19.493 [2024-10-17 17:56:27.390409] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:19.493 [2024-10-17 17:56:27.390416] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:19.493 [2024-10-17 17:56:27.390423] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:19.493 [2024-10-17 17:56:27.391224] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@864 -- # return 0 00:35:20.558 17:56:28 nvmf_dif -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@730 -- # xtrace_disable 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:20.558 17:56:28 nvmf_dif -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:20.558 17:56:28 nvmf_dif -- target/dif.sh@139 -- # create_transport 00:35:20.558 17:56:28 nvmf_dif -- target/dif.sh@50 -- # rpc_cmd nvmf_create_transport -t tcp -o --dif-insert-or-strip 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:20.558 [2024-10-17 17:56:28.129178] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:20.558 17:56:28 nvmf_dif -- target/dif.sh@141 -- # run_test fio_dif_1_default fio_dif_1 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:20.558 17:56:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:20.558 ************************************ 00:35:20.558 START TEST fio_dif_1_default 00:35:20.558 ************************************ 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1125 -- # fio_dif_1 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@86 -- # create_subsystems 0 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@28 -- # local sub 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@30 -- # for sub in "$@" 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@31 -- # create_subsystem 0 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@18 -- # local sub_id=0 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:35:20.558 bdev_null0 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:35:20.558 [2024-10-17 17:56:28.217631] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # fio /dev/fd/62 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # create_json_sub_conf 0 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@558 -- # config=() 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@558 -- # local subsystem config 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:35:20.558 { 00:35:20.558 "params": { 00:35:20.558 "name": "Nvme$subsystem", 00:35:20.558 "trtype": "$TEST_TRANSPORT", 00:35:20.558 "traddr": "$NVMF_FIRST_TARGET_IP", 00:35:20.558 "adrfam": "ipv4", 00:35:20.558 "trsvcid": "$NVMF_PORT", 00:35:20.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:35:20.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:35:20.558 "hdgst": ${hdgst:-false}, 00:35:20.558 "ddgst": ${ddgst:-false} 00:35:20.558 }, 00:35:20.558 "method": "bdev_nvme_attach_controller" 00:35:20.558 } 00:35:20.558 EOF 00:35:20.558 )") 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # gen_fio_conf 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@54 -- # local file 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1339 -- # local sanitizers 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@56 -- # cat 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1341 -- # shift 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1343 -- # local asan_lib= 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@580 -- # cat 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file = 1 )) 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # grep libasan 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file <= files )) 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@582 -- # jq . 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@583 -- # IFS=, 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:35:20.558 "params": { 00:35:20.558 "name": "Nvme0", 00:35:20.558 "trtype": "tcp", 00:35:20.558 "traddr": "10.0.0.2", 00:35:20.558 "adrfam": "ipv4", 00:35:20.558 "trsvcid": "4420", 00:35:20.558 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:35:20.558 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:35:20.558 "hdgst": false, 00:35:20.558 "ddgst": false 00:35:20.558 }, 00:35:20.558 "method": "bdev_nvme_attach_controller" 00:35:20.558 }' 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:35:20.558 17:56:28 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:20.819 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:35:20.819 fio-3.35 00:35:20.819 Starting 1 thread 00:35:33.053 00:35:33.053 filename0: (groupid=0, jobs=1): err= 0: pid=347449: Thu Oct 17 17:56:39 2024 00:35:33.053 read: IOPS=97, BW=389KiB/s (398kB/s)(3904KiB/10036msec) 00:35:33.053 slat (nsec): min=5634, max=38706, avg=6553.08, stdev=1959.80 00:35:33.053 clat (usec): min=40833, max=44185, avg=41109.95, stdev=364.86 00:35:33.053 lat (usec): min=40841, max=44220, avg=41116.50, stdev=365.96 00:35:33.053 clat percentiles (usec): 00:35:33.053 | 1.00th=[40633], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:35:33.053 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:35:33.053 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41681], 95.00th=[42206], 00:35:33.053 | 99.00th=[42206], 99.50th=[42206], 99.90th=[44303], 99.95th=[44303], 00:35:33.053 | 99.99th=[44303] 00:35:33.053 bw ( KiB/s): min= 352, max= 416, per=99.74%, avg=388.80, stdev=15.66, samples=20 00:35:33.053 iops : min= 88, max= 104, avg=97.20, stdev= 3.91, samples=20 00:35:33.053 lat (msec) : 50=100.00% 00:35:33.053 cpu : usr=93.57%, sys=6.19%, ctx=14, majf=0, minf=231 00:35:33.053 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:33.053 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:33.053 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:33.053 issued rwts: total=976,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:33.053 latency : target=0, window=0, percentile=100.00%, depth=4 00:35:33.053 00:35:33.053 Run status group 0 (all jobs): 00:35:33.053 READ: bw=389KiB/s (398kB/s), 389KiB/s-389KiB/s (398kB/s-398kB/s), io=3904KiB (3998kB), run=10036-10036msec 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@88 -- # destroy_subsystems 0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@43 -- # local sub 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@45 -- # for sub in "$@" 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@46 -- # destroy_subsystem 0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@36 -- # local sub_id=0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 00:35:33.053 real 0m11.155s 00:35:33.053 user 0m16.720s 00:35:33.053 sys 0m1.036s 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1126 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 ************************************ 00:35:33.053 END TEST fio_dif_1_default 00:35:33.053 ************************************ 00:35:33.053 17:56:39 nvmf_dif -- target/dif.sh@142 -- # run_test fio_dif_1_multi_subsystems fio_dif_1_multi_subsystems 00:35:33.053 17:56:39 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:35:33.053 17:56:39 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 ************************************ 00:35:33.053 START TEST fio_dif_1_multi_subsystems 00:35:33.053 ************************************ 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1125 -- # fio_dif_1_multi_subsystems 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@92 -- # local files=1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@94 -- # create_subsystems 0 1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@28 -- # local sub 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 bdev_null0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 [2024-10-17 17:56:39.453832] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 bdev_null1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # fio /dev/fd/62 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # create_json_sub_conf 0 1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@558 -- # config=() 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@558 -- # local subsystem config 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:35:33.053 { 00:35:33.053 "params": { 00:35:33.053 "name": "Nvme$subsystem", 00:35:33.053 "trtype": "$TEST_TRANSPORT", 00:35:33.053 "traddr": "$NVMF_FIRST_TARGET_IP", 00:35:33.053 "adrfam": "ipv4", 00:35:33.053 "trsvcid": "$NVMF_PORT", 00:35:33.053 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:35:33.053 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:35:33.053 "hdgst": ${hdgst:-false}, 00:35:33.053 "ddgst": ${ddgst:-false} 00:35:33.053 }, 00:35:33.053 "method": "bdev_nvme_attach_controller" 00:35:33.053 } 00:35:33.053 EOF 00:35:33.053 )") 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # gen_fio_conf 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@54 -- # local file 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1339 -- # local sanitizers 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@56 -- # cat 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:33.053 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1341 -- # shift 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1343 -- # local asan_lib= 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # cat 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file = 1 )) 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # grep libasan 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@73 -- # cat 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:35:33.054 { 00:35:33.054 "params": { 00:35:33.054 "name": "Nvme$subsystem", 00:35:33.054 "trtype": "$TEST_TRANSPORT", 00:35:33.054 "traddr": "$NVMF_FIRST_TARGET_IP", 00:35:33.054 "adrfam": "ipv4", 00:35:33.054 "trsvcid": "$NVMF_PORT", 00:35:33.054 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:35:33.054 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:35:33.054 "hdgst": ${hdgst:-false}, 00:35:33.054 "ddgst": ${ddgst:-false} 00:35:33.054 }, 00:35:33.054 "method": "bdev_nvme_attach_controller" 00:35:33.054 } 00:35:33.054 EOF 00:35:33.054 )") 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file++ )) 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # cat 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # jq . 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@583 -- # IFS=, 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:35:33.054 "params": { 00:35:33.054 "name": "Nvme0", 00:35:33.054 "trtype": "tcp", 00:35:33.054 "traddr": "10.0.0.2", 00:35:33.054 "adrfam": "ipv4", 00:35:33.054 "trsvcid": "4420", 00:35:33.054 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:35:33.054 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:35:33.054 "hdgst": false, 00:35:33.054 "ddgst": false 00:35:33.054 }, 00:35:33.054 "method": "bdev_nvme_attach_controller" 00:35:33.054 },{ 00:35:33.054 "params": { 00:35:33.054 "name": "Nvme1", 00:35:33.054 "trtype": "tcp", 00:35:33.054 "traddr": "10.0.0.2", 00:35:33.054 "adrfam": "ipv4", 00:35:33.054 "trsvcid": "4420", 00:35:33.054 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:35:33.054 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:35:33.054 "hdgst": false, 00:35:33.054 "ddgst": false 00:35:33.054 }, 00:35:33.054 "method": "bdev_nvme_attach_controller" 00:35:33.054 }' 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:35:33.054 17:56:39 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:33.054 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:35:33.054 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:35:33.054 fio-3.35 00:35:33.054 Starting 2 threads 00:35:43.051 00:35:43.051 filename0: (groupid=0, jobs=1): err= 0: pid=349655: Thu Oct 17 17:56:50 2024 00:35:43.051 read: IOPS=189, BW=759KiB/s (777kB/s)(7616KiB/10036msec) 00:35:43.051 slat (nsec): min=5648, max=32088, avg=6625.78, stdev=1446.51 00:35:43.051 clat (usec): min=486, max=41945, avg=21065.92, stdev=20311.14 00:35:43.051 lat (usec): min=492, max=41977, avg=21072.55, stdev=20311.11 00:35:43.051 clat percentiles (usec): 00:35:43.051 | 1.00th=[ 510], 5.00th=[ 652], 10.00th=[ 660], 20.00th=[ 668], 00:35:43.051 | 30.00th=[ 685], 40.00th=[ 709], 50.00th=[40633], 60.00th=[41157], 00:35:43.051 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:35:43.051 | 99.00th=[41681], 99.50th=[41681], 99.90th=[41681], 99.95th=[42206], 00:35:43.051 | 99.99th=[42206] 00:35:43.051 bw ( KiB/s): min= 672, max= 768, per=50.04%, avg=760.00, stdev=25.16, samples=20 00:35:43.051 iops : min= 168, max= 192, avg=190.00, stdev= 6.29, samples=20 00:35:43.051 lat (usec) : 500=0.47%, 750=47.95%, 1000=1.37% 00:35:43.051 lat (msec) : 50=50.21% 00:35:43.051 cpu : usr=96.06%, sys=3.75%, ctx=14, majf=0, minf=135 00:35:43.051 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:43.051 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:43.051 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:43.051 issued rwts: total=1904,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:43.051 latency : target=0, window=0, percentile=100.00%, depth=4 00:35:43.051 filename1: (groupid=0, jobs=1): err= 0: pid=349657: Thu Oct 17 17:56:50 2024 00:35:43.051 read: IOPS=190, BW=760KiB/s (778kB/s)(7632KiB/10040msec) 00:35:43.051 slat (nsec): min=5661, max=31870, avg=6545.40, stdev=1339.11 00:35:43.051 clat (usec): min=553, max=42386, avg=21029.05, stdev=20162.15 00:35:43.051 lat (usec): min=562, max=42392, avg=21035.59, stdev=20162.14 00:35:43.051 clat percentiles (usec): 00:35:43.051 | 1.00th=[ 619], 5.00th=[ 791], 10.00th=[ 816], 20.00th=[ 832], 00:35:43.051 | 30.00th=[ 840], 40.00th=[ 865], 50.00th=[41157], 60.00th=[41157], 00:35:43.051 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:35:43.051 | 99.00th=[41157], 99.50th=[41681], 99.90th=[42206], 99.95th=[42206], 00:35:43.051 | 99.99th=[42206] 00:35:43.051 bw ( KiB/s): min= 704, max= 768, per=50.11%, avg=761.60, stdev=19.70, samples=20 00:35:43.051 iops : min= 176, max= 192, avg=190.40, stdev= 4.92, samples=20 00:35:43.051 lat (usec) : 750=2.31%, 1000=47.59% 00:35:43.051 lat (msec) : 50=50.10% 00:35:43.051 cpu : usr=96.19%, sys=3.61%, ctx=15, majf=0, minf=143 00:35:43.051 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:43.051 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:43.051 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:43.051 issued rwts: total=1908,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:43.051 latency : target=0, window=0, percentile=100.00%, depth=4 00:35:43.051 00:35:43.051 Run status group 0 (all jobs): 00:35:43.051 READ: bw=1519KiB/s (1555kB/s), 759KiB/s-760KiB/s (777kB/s-778kB/s), io=14.9MiB (15.6MB), run=10036-10040msec 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@96 -- # destroy_subsystems 0 1 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@43 -- # local sub 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 0 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=0 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 1 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=1 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.051 00:35:43.051 real 0m11.443s 00:35:43.051 user 0m34.173s 00:35:43.051 sys 0m1.113s 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1126 -- # xtrace_disable 00:35:43.051 17:56:50 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:35:43.051 ************************************ 00:35:43.052 END TEST fio_dif_1_multi_subsystems 00:35:43.052 ************************************ 00:35:43.052 17:56:50 nvmf_dif -- target/dif.sh@143 -- # run_test fio_dif_rand_params fio_dif_rand_params 00:35:43.052 17:56:50 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:35:43.052 17:56:50 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:43.052 17:56:50 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:35:43.052 ************************************ 00:35:43.052 START TEST fio_dif_rand_params 00:35:43.052 ************************************ 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1125 -- # fio_dif_rand_params 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@100 -- # local NULL_DIF 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@101 -- # local bs numjobs runtime iodepth files 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # NULL_DIF=3 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # bs=128k 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # numjobs=3 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # iodepth=3 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # runtime=5 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@105 -- # create_subsystems 0 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:43.052 bdev_null0 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.052 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:43.312 [2024-10-17 17:56:50.981008] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # fio /dev/fd/62 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # create_json_sub_conf 0 00:35:43.312 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # config=() 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # local subsystem config 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:35:43.313 { 00:35:43.313 "params": { 00:35:43.313 "name": "Nvme$subsystem", 00:35:43.313 "trtype": "$TEST_TRANSPORT", 00:35:43.313 "traddr": "$NVMF_FIRST_TARGET_IP", 00:35:43.313 "adrfam": "ipv4", 00:35:43.313 "trsvcid": "$NVMF_PORT", 00:35:43.313 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:35:43.313 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:35:43.313 "hdgst": ${hdgst:-false}, 00:35:43.313 "ddgst": ${ddgst:-false} 00:35:43.313 }, 00:35:43.313 "method": "bdev_nvme_attach_controller" 00:35:43.313 } 00:35:43.313 EOF 00:35:43.313 )") 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # jq . 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@583 -- # IFS=, 00:35:43.313 17:56:50 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:35:43.313 "params": { 00:35:43.313 "name": "Nvme0", 00:35:43.313 "trtype": "tcp", 00:35:43.313 "traddr": "10.0.0.2", 00:35:43.313 "adrfam": "ipv4", 00:35:43.313 "trsvcid": "4420", 00:35:43.313 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:35:43.313 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:35:43.313 "hdgst": false, 00:35:43.313 "ddgst": false 00:35:43.313 }, 00:35:43.313 "method": "bdev_nvme_attach_controller" 00:35:43.313 }' 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:35:43.313 17:56:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:43.573 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:35:43.573 ... 00:35:43.573 fio-3.35 00:35:43.573 Starting 3 threads 00:35:50.166 00:35:50.166 filename0: (groupid=0, jobs=1): err= 0: pid=352094: Thu Oct 17 17:56:57 2024 00:35:50.166 read: IOPS=285, BW=35.7MiB/s (37.5MB/s)(180MiB/5046msec) 00:35:50.166 slat (nsec): min=5661, max=36941, avg=7573.81, stdev=1659.31 00:35:50.166 clat (usec): min=4200, max=90071, avg=10450.22, stdev=11353.74 00:35:50.166 lat (usec): min=4207, max=90079, avg=10457.80, stdev=11353.98 00:35:50.166 clat percentiles (usec): 00:35:50.166 | 1.00th=[ 4686], 5.00th=[ 5473], 10.00th=[ 5932], 20.00th=[ 6456], 00:35:50.166 | 30.00th=[ 6849], 40.00th=[ 7111], 50.00th=[ 7504], 60.00th=[ 7963], 00:35:50.166 | 70.00th=[ 8455], 80.00th=[ 9110], 90.00th=[10159], 95.00th=[46924], 00:35:50.166 | 99.00th=[50070], 99.50th=[86508], 99.90th=[89654], 99.95th=[89654], 00:35:50.166 | 99.99th=[89654] 00:35:50.166 bw ( KiB/s): min=24832, max=51712, per=29.97%, avg=36889.60, stdev=9533.26, samples=10 00:35:50.166 iops : min= 194, max= 404, avg=288.20, stdev=74.48, samples=10 00:35:50.166 lat (msec) : 10=88.57%, 20=4.78%, 50=5.54%, 100=1.11% 00:35:50.166 cpu : usr=94.91%, sys=4.86%, ctx=10, majf=0, minf=102 00:35:50.166 IO depths : 1=0.2%, 2=99.8%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:50.167 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:50.167 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:50.167 issued rwts: total=1443,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:50.167 latency : target=0, window=0, percentile=100.00%, depth=3 00:35:50.167 filename0: (groupid=0, jobs=1): err= 0: pid=352095: Thu Oct 17 17:56:57 2024 00:35:50.167 read: IOPS=346, BW=43.3MiB/s (45.4MB/s)(217MiB/5005msec) 00:35:50.167 slat (nsec): min=5685, max=38560, avg=7838.71, stdev=1760.60 00:35:50.167 clat (usec): min=3733, max=50737, avg=8657.50, stdev=3766.13 00:35:50.167 lat (usec): min=3739, max=50743, avg=8665.34, stdev=3766.24 00:35:50.167 clat percentiles (usec): 00:35:50.167 | 1.00th=[ 4948], 5.00th=[ 5866], 10.00th=[ 6325], 20.00th=[ 6915], 00:35:50.167 | 30.00th=[ 7439], 40.00th=[ 7898], 50.00th=[ 8455], 60.00th=[ 8979], 00:35:50.167 | 70.00th=[ 9372], 80.00th=[ 9765], 90.00th=[10421], 95.00th=[10945], 00:35:50.167 | 99.00th=[12256], 99.50th=[49546], 99.90th=[50594], 99.95th=[50594], 00:35:50.167 | 99.99th=[50594] 00:35:50.167 bw ( KiB/s): min=33024, max=48128, per=35.98%, avg=44288.00, stdev=4456.98, samples=10 00:35:50.167 iops : min= 258, max= 376, avg=346.00, stdev=34.82, samples=10 00:35:50.167 lat (msec) : 4=0.12%, 10=83.66%, 20=15.53%, 50=0.29%, 100=0.40% 00:35:50.167 cpu : usr=93.11%, sys=6.67%, ctx=12, majf=0, minf=130 00:35:50.167 IO depths : 1=0.3%, 2=99.7%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:50.167 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:50.167 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:50.167 issued rwts: total=1732,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:50.167 latency : target=0, window=0, percentile=100.00%, depth=3 00:35:50.167 filename0: (groupid=0, jobs=1): err= 0: pid=352096: Thu Oct 17 17:56:57 2024 00:35:50.167 read: IOPS=332, BW=41.6MiB/s (43.6MB/s)(210MiB/5045msec) 00:35:50.167 slat (nsec): min=5689, max=47440, avg=7693.19, stdev=2341.56 00:35:50.167 clat (usec): min=4612, max=87937, avg=8990.28, stdev=5989.15 00:35:50.167 lat (usec): min=4621, max=87946, avg=8997.97, stdev=5989.27 00:35:50.167 clat percentiles (usec): 00:35:50.167 | 1.00th=[ 5342], 5.00th=[ 5800], 10.00th=[ 6325], 20.00th=[ 6980], 00:35:50.167 | 30.00th=[ 7373], 40.00th=[ 7767], 50.00th=[ 8291], 60.00th=[ 8717], 00:35:50.167 | 70.00th=[ 9110], 80.00th=[ 9503], 90.00th=[10028], 95.00th=[10421], 00:35:50.167 | 99.00th=[47449], 99.50th=[49021], 99.90th=[52167], 99.95th=[87557], 00:35:50.167 | 99.99th=[87557] 00:35:50.167 bw ( KiB/s): min=35911, max=48128, per=34.82%, avg=42861.50, stdev=3861.06, samples=10 00:35:50.167 iops : min= 280, max= 376, avg=334.80, stdev=30.28, samples=10 00:35:50.167 lat (msec) : 10=90.52%, 20=7.45%, 50=1.79%, 100=0.24% 00:35:50.167 cpu : usr=93.62%, sys=6.15%, ctx=13, majf=0, minf=152 00:35:50.167 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:35:50.167 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:50.167 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:35:50.167 issued rwts: total=1677,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:35:50.167 latency : target=0, window=0, percentile=100.00%, depth=3 00:35:50.167 00:35:50.167 Run status group 0 (all jobs): 00:35:50.167 READ: bw=120MiB/s (126MB/s), 35.7MiB/s-43.3MiB/s (37.5MB/s-45.4MB/s), io=607MiB (636MB), run=5005-5046msec 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@107 -- # destroy_subsystems 0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # NULL_DIF=2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # bs=4k 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # numjobs=8 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # iodepth=16 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # runtime= 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # files=2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@111 -- # create_subsystems 0 1 2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 bdev_null0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 [2024-10-17 17:56:57.244277] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 bdev_null1 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null2 64 512 --md-size 16 --dif-type 2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 bdev_null2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 --serial-number 53313233-2 --allow-any-host 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 bdev_null2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # fio /dev/fd/62 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # create_json_sub_conf 0 1 2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 2 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # config=() 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:50.167 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # local subsystem config 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:35:50.168 { 00:35:50.168 "params": { 00:35:50.168 "name": "Nvme$subsystem", 00:35:50.168 "trtype": "$TEST_TRANSPORT", 00:35:50.168 "traddr": "$NVMF_FIRST_TARGET_IP", 00:35:50.168 "adrfam": "ipv4", 00:35:50.168 "trsvcid": "$NVMF_PORT", 00:35:50.168 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:35:50.168 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:35:50.168 "hdgst": ${hdgst:-false}, 00:35:50.168 "ddgst": ${ddgst:-false} 00:35:50.168 }, 00:35:50.168 "method": "bdev_nvme_attach_controller" 00:35:50.168 } 00:35:50.168 EOF 00:35:50.168 )") 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:35:50.168 { 00:35:50.168 "params": { 00:35:50.168 "name": "Nvme$subsystem", 00:35:50.168 "trtype": "$TEST_TRANSPORT", 00:35:50.168 "traddr": "$NVMF_FIRST_TARGET_IP", 00:35:50.168 "adrfam": "ipv4", 00:35:50.168 "trsvcid": "$NVMF_PORT", 00:35:50.168 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:35:50.168 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:35:50.168 "hdgst": ${hdgst:-false}, 00:35:50.168 "ddgst": ${ddgst:-false} 00:35:50.168 }, 00:35:50.168 "method": "bdev_nvme_attach_controller" 00:35:50.168 } 00:35:50.168 EOF 00:35:50.168 )") 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:35:50.168 { 00:35:50.168 "params": { 00:35:50.168 "name": "Nvme$subsystem", 00:35:50.168 "trtype": "$TEST_TRANSPORT", 00:35:50.168 "traddr": "$NVMF_FIRST_TARGET_IP", 00:35:50.168 "adrfam": "ipv4", 00:35:50.168 "trsvcid": "$NVMF_PORT", 00:35:50.168 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:35:50.168 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:35:50.168 "hdgst": ${hdgst:-false}, 00:35:50.168 "ddgst": ${ddgst:-false} 00:35:50.168 }, 00:35:50.168 "method": "bdev_nvme_attach_controller" 00:35:50.168 } 00:35:50.168 EOF 00:35:50.168 )") 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # jq . 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@583 -- # IFS=, 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:35:50.168 "params": { 00:35:50.168 "name": "Nvme0", 00:35:50.168 "trtype": "tcp", 00:35:50.168 "traddr": "10.0.0.2", 00:35:50.168 "adrfam": "ipv4", 00:35:50.168 "trsvcid": "4420", 00:35:50.168 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:35:50.168 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:35:50.168 "hdgst": false, 00:35:50.168 "ddgst": false 00:35:50.168 }, 00:35:50.168 "method": "bdev_nvme_attach_controller" 00:35:50.168 },{ 00:35:50.168 "params": { 00:35:50.168 "name": "Nvme1", 00:35:50.168 "trtype": "tcp", 00:35:50.168 "traddr": "10.0.0.2", 00:35:50.168 "adrfam": "ipv4", 00:35:50.168 "trsvcid": "4420", 00:35:50.168 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:35:50.168 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:35:50.168 "hdgst": false, 00:35:50.168 "ddgst": false 00:35:50.168 }, 00:35:50.168 "method": "bdev_nvme_attach_controller" 00:35:50.168 },{ 00:35:50.168 "params": { 00:35:50.168 "name": "Nvme2", 00:35:50.168 "trtype": "tcp", 00:35:50.168 "traddr": "10.0.0.2", 00:35:50.168 "adrfam": "ipv4", 00:35:50.168 "trsvcid": "4420", 00:35:50.168 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:35:50.168 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:35:50.168 "hdgst": false, 00:35:50.168 "ddgst": false 00:35:50.168 }, 00:35:50.168 "method": "bdev_nvme_attach_controller" 00:35:50.168 }' 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:35:50.168 17:56:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:35:50.168 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:35:50.168 ... 00:35:50.168 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:35:50.168 ... 00:35:50.168 filename2: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:35:50.168 ... 00:35:50.168 fio-3.35 00:35:50.168 Starting 24 threads 00:36:02.375 00:36:02.375 filename0: (groupid=0, jobs=1): err= 0: pid=353346: Thu Oct 17 17:57:08 2024 00:36:02.375 read: IOPS=749, BW=2997KiB/s (3069kB/s)(29.3MiB/10006msec) 00:36:02.375 slat (usec): min=5, max=137, avg=13.89, stdev=13.19 00:36:02.375 clat (usec): min=1535, max=38393, avg=21247.17, stdev=4640.55 00:36:02.375 lat (usec): min=1553, max=38403, avg=21261.06, stdev=4642.19 00:36:02.375 clat percentiles (usec): 00:36:02.375 | 1.00th=[ 3261], 5.00th=[13698], 10.00th=[14615], 20.00th=[18220], 00:36:02.375 | 30.00th=[22152], 40.00th=[22676], 50.00th=[22938], 60.00th=[23200], 00:36:02.375 | 70.00th=[23200], 80.00th=[23462], 90.00th=[24249], 95.00th=[24773], 00:36:02.375 | 99.00th=[33424], 99.50th=[34866], 99.90th=[37487], 99.95th=[38536], 00:36:02.375 | 99.99th=[38536] 00:36:02.375 bw ( KiB/s): min= 2560, max= 4168, per=4.49%, avg=3008.42, stdev=445.07, samples=19 00:36:02.375 iops : min= 640, max= 1042, avg=752.11, stdev=111.27, samples=19 00:36:02.375 lat (msec) : 2=0.33%, 4=1.32%, 10=1.37%, 20=22.61%, 50=74.36% 00:36:02.375 cpu : usr=97.95%, sys=1.19%, ctx=278, majf=0, minf=39 00:36:02.375 IO depths : 1=4.2%, 2=8.4%, 4=18.8%, 8=60.1%, 16=8.5%, 32=0.0%, >=64=0.0% 00:36:02.375 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 complete : 0=0.0%, 4=92.4%, 8=2.0%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 issued rwts: total=7497,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.375 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.375 filename0: (groupid=0, jobs=1): err= 0: pid=353347: Thu Oct 17 17:57:08 2024 00:36:02.375 read: IOPS=688, BW=2756KiB/s (2822kB/s)(26.9MiB/10008msec) 00:36:02.375 slat (usec): min=5, max=122, avg=27.89, stdev=20.93 00:36:02.375 clat (usec): min=10009, max=46228, avg=22977.95, stdev=2887.67 00:36:02.375 lat (usec): min=10015, max=46246, avg=23005.84, stdev=2890.01 00:36:02.375 clat percentiles (usec): 00:36:02.375 | 1.00th=[14091], 5.00th=[17433], 10.00th=[21103], 20.00th=[22414], 00:36:02.375 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.375 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24511], 95.00th=[26084], 00:36:02.375 | 99.00th=[33817], 99.50th=[34341], 99.90th=[39060], 99.95th=[45876], 00:36:02.375 | 99.99th=[46400] 00:36:02.375 bw ( KiB/s): min= 2560, max= 3008, per=4.12%, avg=2758.74, stdev=110.63, samples=19 00:36:02.375 iops : min= 640, max= 752, avg=689.68, stdev=27.66, samples=19 00:36:02.375 lat (msec) : 20=8.73%, 50=91.27% 00:36:02.375 cpu : usr=99.06%, sys=0.65%, ctx=16, majf=0, minf=32 00:36:02.375 IO depths : 1=4.8%, 2=9.5%, 4=20.3%, 8=57.5%, 16=8.0%, 32=0.0%, >=64=0.0% 00:36:02.375 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 complete : 0=0.0%, 4=92.7%, 8=1.7%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 issued rwts: total=6895,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.375 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.375 filename0: (groupid=0, jobs=1): err= 0: pid=353348: Thu Oct 17 17:57:08 2024 00:36:02.375 read: IOPS=704, BW=2816KiB/s (2884kB/s)(27.5MiB/10005msec) 00:36:02.375 slat (usec): min=5, max=459, avg=31.61, stdev=25.11 00:36:02.375 clat (usec): min=5740, max=38339, avg=22446.21, stdev=3012.86 00:36:02.375 lat (usec): min=5745, max=38364, avg=22477.82, stdev=3016.54 00:36:02.375 clat percentiles (usec): 00:36:02.375 | 1.00th=[13698], 5.00th=[16188], 10.00th=[18482], 20.00th=[22152], 00:36:02.375 | 30.00th=[22414], 40.00th=[22676], 50.00th=[22938], 60.00th=[22938], 00:36:02.375 | 70.00th=[23200], 80.00th=[23462], 90.00th=[24249], 95.00th=[25035], 00:36:02.375 | 99.00th=[31851], 99.50th=[33817], 99.90th=[38536], 99.95th=[38536], 00:36:02.375 | 99.99th=[38536] 00:36:02.375 bw ( KiB/s): min= 2560, max= 3024, per=4.20%, avg=2810.95, stdev=127.00, samples=19 00:36:02.375 iops : min= 640, max= 756, avg=702.74, stdev=31.75, samples=19 00:36:02.375 lat (msec) : 10=0.45%, 20=12.82%, 50=86.73% 00:36:02.375 cpu : usr=98.81%, sys=0.89%, ctx=20, majf=0, minf=25 00:36:02.375 IO depths : 1=4.3%, 2=8.7%, 4=18.8%, 8=59.5%, 16=8.7%, 32=0.0%, >=64=0.0% 00:36:02.375 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 complete : 0=0.0%, 4=92.4%, 8=2.3%, 16=5.3%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 issued rwts: total=7044,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.375 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.375 filename0: (groupid=0, jobs=1): err= 0: pid=353349: Thu Oct 17 17:57:08 2024 00:36:02.375 read: IOPS=695, BW=2784KiB/s (2851kB/s)(27.2MiB/10014msec) 00:36:02.375 slat (usec): min=5, max=143, avg=31.53, stdev=25.46 00:36:02.375 clat (usec): min=6642, max=39506, avg=22701.07, stdev=2807.51 00:36:02.375 lat (usec): min=6652, max=39552, avg=22732.60, stdev=2810.62 00:36:02.375 clat percentiles (usec): 00:36:02.375 | 1.00th=[13304], 5.00th=[16712], 10.00th=[20055], 20.00th=[22414], 00:36:02.375 | 30.00th=[22676], 40.00th=[22676], 50.00th=[22938], 60.00th=[23200], 00:36:02.375 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[25035], 00:36:02.375 | 99.00th=[32637], 99.50th=[36439], 99.90th=[39060], 99.95th=[39584], 00:36:02.375 | 99.99th=[39584] 00:36:02.375 bw ( KiB/s): min= 2688, max= 2992, per=4.15%, avg=2781.45, stdev=96.21, samples=20 00:36:02.375 iops : min= 672, max= 748, avg=695.35, stdev=24.04, samples=20 00:36:02.375 lat (msec) : 10=0.17%, 20=9.47%, 50=90.36% 00:36:02.375 cpu : usr=98.67%, sys=0.92%, ctx=48, majf=0, minf=40 00:36:02.375 IO depths : 1=4.8%, 2=9.7%, 4=20.7%, 8=56.8%, 16=8.0%, 32=0.0%, >=64=0.0% 00:36:02.375 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 complete : 0=0.0%, 4=93.0%, 8=1.5%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.375 issued rwts: total=6969,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.375 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.375 filename0: (groupid=0, jobs=1): err= 0: pid=353350: Thu Oct 17 17:57:08 2024 00:36:02.375 read: IOPS=740, BW=2961KiB/s (3032kB/s)(29.0MiB/10018msec) 00:36:02.375 slat (usec): min=5, max=131, avg=13.25, stdev=13.22 00:36:02.375 clat (usec): min=2665, max=42417, avg=21520.81, stdev=4500.94 00:36:02.375 lat (usec): min=2677, max=42446, avg=21534.05, stdev=4502.26 00:36:02.375 clat percentiles (usec): 00:36:02.376 | 1.00th=[ 3589], 5.00th=[14091], 10.00th=[15664], 20.00th=[17957], 00:36:02.376 | 30.00th=[20841], 40.00th=[22676], 50.00th=[22938], 60.00th=[23200], 00:36:02.376 | 70.00th=[23200], 80.00th=[23725], 90.00th=[24511], 95.00th=[26870], 00:36:02.376 | 99.00th=[33817], 99.50th=[35914], 99.90th=[40109], 99.95th=[42206], 00:36:02.376 | 99.99th=[42206] 00:36:02.376 bw ( KiB/s): min= 2688, max= 4224, per=4.42%, avg=2962.40, stdev=355.74, samples=20 00:36:02.376 iops : min= 672, max= 1056, avg=740.60, stdev=88.94, samples=20 00:36:02.376 lat (msec) : 4=1.09%, 10=0.63%, 20=25.81%, 50=72.46% 00:36:02.376 cpu : usr=98.62%, sys=0.95%, ctx=185, majf=0, minf=45 00:36:02.376 IO depths : 1=1.2%, 2=2.5%, 4=9.6%, 8=74.8%, 16=11.9%, 32=0.0%, >=64=0.0% 00:36:02.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 complete : 0=0.0%, 4=90.0%, 8=5.0%, 16=5.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 issued rwts: total=7416,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.376 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.376 filename0: (groupid=0, jobs=1): err= 0: pid=353351: Thu Oct 17 17:57:08 2024 00:36:02.376 read: IOPS=691, BW=2764KiB/s (2831kB/s)(27.0MiB/10004msec) 00:36:02.376 slat (usec): min=5, max=145, avg=35.77, stdev=23.47 00:36:02.376 clat (usec): min=3786, max=52285, avg=22808.14, stdev=2848.85 00:36:02.376 lat (usec): min=3792, max=52304, avg=22843.91, stdev=2851.62 00:36:02.376 clat percentiles (usec): 00:36:02.376 | 1.00th=[13435], 5.00th=[19006], 10.00th=[22152], 20.00th=[22414], 00:36:02.376 | 30.00th=[22676], 40.00th=[22676], 50.00th=[22938], 60.00th=[23200], 00:36:02.376 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24773], 00:36:02.376 | 99.00th=[29492], 99.50th=[34341], 99.90th=[52167], 99.95th=[52167], 00:36:02.376 | 99.99th=[52167] 00:36:02.376 bw ( KiB/s): min= 2533, max= 3104, per=4.10%, avg=2747.21, stdev=122.65, samples=19 00:36:02.376 iops : min= 633, max= 776, avg=686.79, stdev=30.69, samples=19 00:36:02.376 lat (msec) : 4=0.06%, 10=0.78%, 20=4.73%, 50=94.20%, 100=0.23% 00:36:02.376 cpu : usr=99.15%, sys=0.55%, ctx=16, majf=0, minf=26 00:36:02.376 IO depths : 1=5.5%, 2=11.1%, 4=22.9%, 8=53.4%, 16=7.1%, 32=0.0%, >=64=0.0% 00:36:02.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 complete : 0=0.0%, 4=93.5%, 8=0.8%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 issued rwts: total=6914,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.376 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.376 filename0: (groupid=0, jobs=1): err= 0: pid=353352: Thu Oct 17 17:57:08 2024 00:36:02.376 read: IOPS=709, BW=2838KiB/s (2906kB/s)(27.8MiB/10017msec) 00:36:02.376 slat (usec): min=5, max=121, avg=21.42, stdev=18.42 00:36:02.376 clat (usec): min=10085, max=38384, avg=22387.03, stdev=3255.62 00:36:02.376 lat (usec): min=10091, max=38396, avg=22408.45, stdev=3258.39 00:36:02.376 clat percentiles (usec): 00:36:02.376 | 1.00th=[13435], 5.00th=[15270], 10.00th=[17433], 20.00th=[22152], 00:36:02.376 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.376 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[25297], 00:36:02.376 | 99.00th=[33817], 99.50th=[34341], 99.90th=[37487], 99.95th=[38536], 00:36:02.376 | 99.99th=[38536] 00:36:02.376 bw ( KiB/s): min= 2656, max= 3280, per=4.24%, avg=2838.40, stdev=182.46, samples=20 00:36:02.376 iops : min= 664, max= 820, avg=709.60, stdev=45.61, samples=20 00:36:02.376 lat (msec) : 20=16.52%, 50=83.48% 00:36:02.376 cpu : usr=98.89%, sys=0.82%, ctx=18, majf=0, minf=47 00:36:02.376 IO depths : 1=4.2%, 2=8.6%, 4=19.1%, 8=59.6%, 16=8.5%, 32=0.0%, >=64=0.0% 00:36:02.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 complete : 0=0.0%, 4=92.5%, 8=2.0%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 issued rwts: total=7106,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.376 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.376 filename0: (groupid=0, jobs=1): err= 0: pid=353353: Thu Oct 17 17:57:08 2024 00:36:02.376 read: IOPS=689, BW=2756KiB/s (2822kB/s)(26.9MiB/10011msec) 00:36:02.376 slat (usec): min=6, max=106, avg=20.51, stdev=15.79 00:36:02.376 clat (usec): min=11731, max=35005, avg=23027.93, stdev=1643.87 00:36:02.376 lat (usec): min=11740, max=35014, avg=23048.44, stdev=1645.21 00:36:02.376 clat percentiles (usec): 00:36:02.376 | 1.00th=[14746], 5.00th=[21890], 10.00th=[22414], 20.00th=[22676], 00:36:02.376 | 30.00th=[22938], 40.00th=[22938], 50.00th=[23200], 60.00th=[23200], 00:36:02.376 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24511], 00:36:02.376 | 99.00th=[25560], 99.50th=[26346], 99.90th=[33162], 99.95th=[34866], 00:36:02.376 | 99.99th=[34866] 00:36:02.376 bw ( KiB/s): min= 2688, max= 3088, per=4.12%, avg=2756.21, stdev=110.32, samples=19 00:36:02.376 iops : min= 672, max= 772, avg=689.05, stdev=27.58, samples=19 00:36:02.376 lat (msec) : 20=3.65%, 50=96.35% 00:36:02.376 cpu : usr=98.76%, sys=0.78%, ctx=85, majf=0, minf=47 00:36:02.376 IO depths : 1=6.0%, 2=12.1%, 4=24.3%, 8=51.1%, 16=6.5%, 32=0.0%, >=64=0.0% 00:36:02.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 complete : 0=0.0%, 4=93.9%, 8=0.2%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 issued rwts: total=6898,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.376 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.376 filename1: (groupid=0, jobs=1): err= 0: pid=353354: Thu Oct 17 17:57:08 2024 00:36:02.376 read: IOPS=711, BW=2846KiB/s (2914kB/s)(27.8MiB/10005msec) 00:36:02.376 slat (usec): min=5, max=455, avg=16.55, stdev=17.01 00:36:02.376 clat (usec): min=5328, max=40110, avg=22387.74, stdev=3870.54 00:36:02.376 lat (usec): min=5346, max=40117, avg=22404.29, stdev=3871.97 00:36:02.376 clat percentiles (usec): 00:36:02.376 | 1.00th=[13304], 5.00th=[14746], 10.00th=[16712], 20.00th=[20317], 00:36:02.376 | 30.00th=[22414], 40.00th=[22676], 50.00th=[22938], 60.00th=[23200], 00:36:02.376 | 70.00th=[23462], 80.00th=[23987], 90.00th=[25560], 95.00th=[29492], 00:36:02.376 | 99.00th=[33424], 99.50th=[34866], 99.90th=[38011], 99.95th=[40109], 00:36:02.376 | 99.99th=[40109] 00:36:02.376 bw ( KiB/s): min= 2560, max= 3152, per=4.25%, avg=2842.11, stdev=137.56, samples=19 00:36:02.376 iops : min= 640, max= 788, avg=710.53, stdev=34.39, samples=19 00:36:02.376 lat (msec) : 10=0.34%, 20=18.88%, 50=80.78% 00:36:02.376 cpu : usr=98.83%, sys=0.87%, ctx=16, majf=0, minf=28 00:36:02.376 IO depths : 1=1.1%, 2=2.2%, 4=7.2%, 8=75.5%, 16=14.0%, 32=0.0%, >=64=0.0% 00:36:02.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 complete : 0=0.0%, 4=89.9%, 8=7.0%, 16=3.1%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 issued rwts: total=7118,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.376 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.376 filename1: (groupid=0, jobs=1): err= 0: pid=353355: Thu Oct 17 17:57:08 2024 00:36:02.376 read: IOPS=697, BW=2789KiB/s (2856kB/s)(27.3MiB/10022msec) 00:36:02.376 slat (usec): min=5, max=113, avg=18.31, stdev=15.08 00:36:02.376 clat (usec): min=2971, max=38116, avg=22792.76, stdev=2872.25 00:36:02.376 lat (usec): min=2988, max=38136, avg=22811.07, stdev=2873.12 00:36:02.376 clat percentiles (usec): 00:36:02.376 | 1.00th=[ 5800], 5.00th=[20841], 10.00th=[22414], 20.00th=[22676], 00:36:02.376 | 30.00th=[22938], 40.00th=[22938], 50.00th=[23200], 60.00th=[23462], 00:36:02.376 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24773], 00:36:02.376 | 99.00th=[25822], 99.50th=[26084], 99.90th=[35390], 99.95th=[38011], 00:36:02.376 | 99.99th=[38011] 00:36:02.376 bw ( KiB/s): min= 2688, max= 3559, per=4.17%, avg=2789.15, stdev=196.58, samples=20 00:36:02.376 iops : min= 672, max= 889, avg=697.25, stdev=48.99, samples=20 00:36:02.376 lat (msec) : 4=0.83%, 10=0.86%, 20=2.98%, 50=95.33% 00:36:02.376 cpu : usr=98.97%, sys=0.73%, ctx=13, majf=0, minf=29 00:36:02.376 IO depths : 1=5.9%, 2=11.9%, 4=24.2%, 8=51.3%, 16=6.6%, 32=0.0%, >=64=0.0% 00:36:02.376 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 complete : 0=0.0%, 4=93.9%, 8=0.2%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.376 issued rwts: total=6988,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.376 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.376 filename1: (groupid=0, jobs=1): err= 0: pid=353356: Thu Oct 17 17:57:08 2024 00:36:02.376 read: IOPS=692, BW=2770KiB/s (2836kB/s)(27.1MiB/10006msec) 00:36:02.376 slat (usec): min=5, max=152, avg=25.44, stdev=23.06 00:36:02.376 clat (usec): min=5195, max=49043, avg=22907.23, stdev=3289.96 00:36:02.376 lat (usec): min=5201, max=49063, avg=22932.67, stdev=3290.37 00:36:02.376 clat percentiles (usec): 00:36:02.376 | 1.00th=[13173], 5.00th=[16581], 10.00th=[20317], 20.00th=[22414], 00:36:02.376 | 30.00th=[22676], 40.00th=[22938], 50.00th=[23200], 60.00th=[23200], 00:36:02.376 | 70.00th=[23462], 80.00th=[23987], 90.00th=[24511], 95.00th=[26084], 00:36:02.376 | 99.00th=[35390], 99.50th=[38536], 99.90th=[42206], 99.95th=[49021], 00:36:02.376 | 99.99th=[49021] 00:36:02.376 bw ( KiB/s): min= 2560, max= 3024, per=4.12%, avg=2759.58, stdev=127.82, samples=19 00:36:02.376 iops : min= 640, max= 756, avg=689.89, stdev=31.95, samples=19 00:36:02.376 lat (msec) : 10=0.59%, 20=8.67%, 50=90.73% 00:36:02.376 cpu : usr=98.98%, sys=0.70%, ctx=32, majf=0, minf=41 00:36:02.376 IO depths : 1=2.5%, 2=5.0%, 4=12.2%, 8=68.0%, 16=12.3%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=90.7%, 8=5.9%, 16=3.4%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=6928,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename1: (groupid=0, jobs=1): err= 0: pid=353357: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=684, BW=2739KiB/s (2804kB/s)(26.8MiB/10017msec) 00:36:02.377 slat (usec): min=5, max=462, avg=32.24, stdev=22.85 00:36:02.377 clat (usec): min=7405, max=39078, avg=23085.41, stdev=1426.08 00:36:02.377 lat (usec): min=7415, max=39092, avg=23117.65, stdev=1426.52 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[16581], 5.00th=[22152], 10.00th=[22414], 20.00th=[22676], 00:36:02.377 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.377 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24511], 00:36:02.377 | 99.00th=[25297], 99.50th=[25560], 99.90th=[35390], 99.95th=[39060], 00:36:02.377 | 99.99th=[39060] 00:36:02.377 bw ( KiB/s): min= 2560, max= 2944, per=4.09%, avg=2736.80, stdev=89.23, samples=20 00:36:02.377 iops : min= 640, max= 736, avg=684.20, stdev=22.31, samples=20 00:36:02.377 lat (msec) : 10=0.13%, 20=1.59%, 50=98.28% 00:36:02.377 cpu : usr=98.96%, sys=0.74%, ctx=18, majf=0, minf=26 00:36:02.377 IO depths : 1=6.2%, 2=12.3%, 4=24.7%, 8=50.5%, 16=6.3%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=94.0%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=6858,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename1: (groupid=0, jobs=1): err= 0: pid=353358: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=698, BW=2793KiB/s (2860kB/s)(27.3MiB/10014msec) 00:36:02.377 slat (usec): min=5, max=125, avg=26.05, stdev=20.30 00:36:02.377 clat (usec): min=9310, max=39117, avg=22699.90, stdev=2798.30 00:36:02.377 lat (usec): min=9324, max=39126, avg=22725.95, stdev=2800.13 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[13566], 5.00th=[16909], 10.00th=[19268], 20.00th=[22414], 00:36:02.377 | 30.00th=[22676], 40.00th=[22676], 50.00th=[22938], 60.00th=[23200], 00:36:02.377 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[25035], 00:36:02.377 | 99.00th=[33424], 99.50th=[35914], 99.90th=[39060], 99.95th=[39060], 00:36:02.377 | 99.99th=[39060] 00:36:02.377 bw ( KiB/s): min= 2688, max= 3216, per=4.17%, avg=2790.40, stdev=136.00, samples=20 00:36:02.377 iops : min= 672, max= 804, avg=697.60, stdev=34.00, samples=20 00:36:02.377 lat (msec) : 10=0.06%, 20=10.67%, 50=89.27% 00:36:02.377 cpu : usr=98.82%, sys=0.88%, ctx=17, majf=0, minf=31 00:36:02.377 IO depths : 1=4.6%, 2=9.3%, 4=21.0%, 8=57.1%, 16=8.0%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=93.1%, 8=1.2%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=6992,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename1: (groupid=0, jobs=1): err= 0: pid=353359: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=687, BW=2749KiB/s (2815kB/s)(26.9MiB/10011msec) 00:36:02.377 slat (usec): min=5, max=124, avg=15.41, stdev=14.25 00:36:02.377 clat (usec): min=6164, max=40400, avg=23147.71, stdev=1728.67 00:36:02.377 lat (usec): min=6171, max=40408, avg=23163.12, stdev=1727.11 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[14353], 5.00th=[22152], 10.00th=[22414], 20.00th=[22676], 00:36:02.377 | 30.00th=[22938], 40.00th=[22938], 50.00th=[23200], 60.00th=[23462], 00:36:02.377 | 70.00th=[23462], 80.00th=[23987], 90.00th=[24249], 95.00th=[24773], 00:36:02.377 | 99.00th=[25822], 99.50th=[26084], 99.90th=[35914], 99.95th=[40633], 00:36:02.377 | 99.99th=[40633] 00:36:02.377 bw ( KiB/s): min= 2560, max= 2944, per=4.11%, avg=2748.63, stdev=103.50, samples=19 00:36:02.377 iops : min= 640, max= 736, avg=687.16, stdev=25.87, samples=19 00:36:02.377 lat (msec) : 10=0.09%, 20=2.35%, 50=97.56% 00:36:02.377 cpu : usr=99.02%, sys=0.69%, ctx=16, majf=0, minf=48 00:36:02.377 IO depths : 1=6.0%, 2=12.2%, 4=24.6%, 8=50.7%, 16=6.5%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=94.0%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=6880,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename1: (groupid=0, jobs=1): err= 0: pid=353360: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=681, BW=2726KiB/s (2792kB/s)(26.6MiB/10001msec) 00:36:02.377 slat (usec): min=5, max=142, avg=35.36, stdev=23.17 00:36:02.377 clat (usec): min=10606, max=33121, avg=23141.28, stdev=1181.64 00:36:02.377 lat (usec): min=10612, max=33136, avg=23176.64, stdev=1179.13 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[21627], 5.00th=[22152], 10.00th=[22414], 20.00th=[22414], 00:36:02.377 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.377 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24511], 00:36:02.377 | 99.00th=[25297], 99.50th=[26084], 99.90th=[32900], 99.95th=[33162], 00:36:02.377 | 99.99th=[33162] 00:36:02.377 bw ( KiB/s): min= 2560, max= 2816, per=4.06%, avg=2721.68, stdev=71.93, samples=19 00:36:02.377 iops : min= 640, max= 704, avg=680.42, stdev=17.98, samples=19 00:36:02.377 lat (msec) : 20=0.47%, 50=99.53% 00:36:02.377 cpu : usr=99.01%, sys=0.69%, ctx=17, majf=0, minf=34 00:36:02.377 IO depths : 1=6.2%, 2=12.4%, 4=24.9%, 8=50.2%, 16=6.3%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=6816,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename1: (groupid=0, jobs=1): err= 0: pid=353361: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=689, BW=2758KiB/s (2824kB/s)(26.9MiB/10004msec) 00:36:02.377 slat (usec): min=5, max=132, avg=25.53, stdev=24.31 00:36:02.377 clat (usec): min=3908, max=41927, avg=23059.23, stdev=3581.83 00:36:02.377 lat (usec): min=3916, max=41961, avg=23084.76, stdev=3582.70 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[13042], 5.00th=[16909], 10.00th=[18744], 20.00th=[22152], 00:36:02.377 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.377 | 70.00th=[23725], 80.00th=[24249], 90.00th=[26084], 95.00th=[29492], 00:36:02.377 | 99.00th=[34341], 99.50th=[38536], 99.90th=[41681], 99.95th=[41681], 00:36:02.377 | 99.99th=[41681] 00:36:02.377 bw ( KiB/s): min= 2549, max= 2896, per=4.10%, avg=2747.21, stdev=78.04, samples=19 00:36:02.377 iops : min= 637, max= 724, avg=686.79, stdev=19.55, samples=19 00:36:02.377 lat (msec) : 4=0.09%, 10=0.46%, 20=12.39%, 50=87.05% 00:36:02.377 cpu : usr=98.70%, sys=0.80%, ctx=95, majf=0, minf=40 00:36:02.377 IO depths : 1=0.9%, 2=1.9%, 4=6.5%, 8=76.2%, 16=14.6%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=89.9%, 8=7.4%, 16=2.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=6898,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename2: (groupid=0, jobs=1): err= 0: pid=353362: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=684, BW=2737KiB/s (2802kB/s)(26.7MiB/10004msec) 00:36:02.377 slat (usec): min=5, max=141, avg=19.88, stdev=19.95 00:36:02.377 clat (usec): min=4121, max=57073, avg=23295.70, stdev=3354.39 00:36:02.377 lat (usec): min=4129, max=57096, avg=23315.58, stdev=3354.27 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[13566], 5.00th=[17695], 10.00th=[21103], 20.00th=[22676], 00:36:02.377 | 30.00th=[22938], 40.00th=[22938], 50.00th=[23200], 60.00th=[23462], 00:36:02.377 | 70.00th=[23725], 80.00th=[24249], 90.00th=[25035], 95.00th=[28705], 00:36:02.377 | 99.00th=[34866], 99.50th=[39060], 99.90th=[44303], 99.95th=[56886], 00:36:02.377 | 99.99th=[56886] 00:36:02.377 bw ( KiB/s): min= 2576, max= 2880, per=4.07%, avg=2725.32, stdev=77.69, samples=19 00:36:02.377 iops : min= 644, max= 720, avg=681.32, stdev=19.45, samples=19 00:36:02.377 lat (msec) : 10=0.45%, 20=7.63%, 50=91.86%, 100=0.06% 00:36:02.377 cpu : usr=98.68%, sys=0.86%, ctx=74, majf=0, minf=28 00:36:02.377 IO depths : 1=0.4%, 2=0.8%, 4=3.3%, 8=79.0%, 16=16.5%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=89.6%, 8=8.8%, 16=1.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=6844,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename2: (groupid=0, jobs=1): err= 0: pid=353363: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=713, BW=2856KiB/s (2924kB/s)(27.9MiB/10017msec) 00:36:02.377 slat (usec): min=5, max=118, avg=17.39, stdev=17.13 00:36:02.377 clat (usec): min=8442, max=40641, avg=22272.55, stdev=3437.07 00:36:02.377 lat (usec): min=8451, max=40658, avg=22289.94, stdev=3439.00 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[12518], 5.00th=[15139], 10.00th=[16712], 20.00th=[21365], 00:36:02.377 | 30.00th=[22414], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.377 | 70.00th=[23462], 80.00th=[23987], 90.00th=[24511], 95.00th=[25560], 00:36:02.377 | 99.00th=[32113], 99.50th=[34341], 99.90th=[40633], 99.95th=[40633], 00:36:02.377 | 99.99th=[40633] 00:36:02.377 bw ( KiB/s): min= 2688, max= 3328, per=4.27%, avg=2856.80, stdev=179.33, samples=20 00:36:02.377 iops : min= 672, max= 832, avg=714.20, stdev=44.83, samples=20 00:36:02.377 lat (msec) : 10=0.49%, 20=16.55%, 50=82.96% 00:36:02.377 cpu : usr=98.78%, sys=0.93%, ctx=22, majf=0, minf=37 00:36:02.377 IO depths : 1=4.0%, 2=8.1%, 4=18.1%, 8=60.9%, 16=8.8%, 32=0.0%, >=64=0.0% 00:36:02.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 complete : 0=0.0%, 4=92.3%, 8=2.3%, 16=5.4%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.377 issued rwts: total=7152,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.377 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.377 filename2: (groupid=0, jobs=1): err= 0: pid=353364: Thu Oct 17 17:57:08 2024 00:36:02.377 read: IOPS=682, BW=2728KiB/s (2794kB/s)(26.7MiB/10016msec) 00:36:02.377 slat (usec): min=5, max=110, avg=16.05, stdev=15.35 00:36:02.377 clat (usec): min=11910, max=37116, avg=23332.18, stdev=1546.45 00:36:02.377 lat (usec): min=11918, max=37191, avg=23348.23, stdev=1547.52 00:36:02.377 clat percentiles (usec): 00:36:02.377 | 1.00th=[16712], 5.00th=[22414], 10.00th=[22676], 20.00th=[22676], 00:36:02.378 | 30.00th=[22938], 40.00th=[23200], 50.00th=[23200], 60.00th=[23462], 00:36:02.378 | 70.00th=[23725], 80.00th=[23987], 90.00th=[24249], 95.00th=[24773], 00:36:02.378 | 99.00th=[29492], 99.50th=[31327], 99.90th=[36963], 99.95th=[36963], 00:36:02.378 | 99.99th=[36963] 00:36:02.378 bw ( KiB/s): min= 2560, max= 2816, per=4.07%, avg=2726.40, stdev=70.54, samples=20 00:36:02.378 iops : min= 640, max= 704, avg=681.60, stdev=17.63, samples=20 00:36:02.378 lat (msec) : 20=1.51%, 50=98.49% 00:36:02.378 cpu : usr=98.91%, sys=0.80%, ctx=16, majf=0, minf=29 00:36:02.378 IO depths : 1=5.0%, 2=10.1%, 4=22.2%, 8=55.0%, 16=7.6%, 32=0.0%, >=64=0.0% 00:36:02.378 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 complete : 0=0.0%, 4=92.9%, 8=1.4%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 issued rwts: total=6832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.378 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.378 filename2: (groupid=0, jobs=1): err= 0: pid=353365: Thu Oct 17 17:57:08 2024 00:36:02.378 read: IOPS=693, BW=2776KiB/s (2842kB/s)(27.1MiB/10015msec) 00:36:02.378 slat (usec): min=5, max=130, avg=28.49, stdev=20.37 00:36:02.378 clat (usec): min=9402, max=39759, avg=22816.07, stdev=2676.02 00:36:02.378 lat (usec): min=9415, max=39768, avg=22844.56, stdev=2678.18 00:36:02.378 clat percentiles (usec): 00:36:02.378 | 1.00th=[13960], 5.00th=[17433], 10.00th=[21627], 20.00th=[22414], 00:36:02.378 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.378 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[25035], 00:36:02.378 | 99.00th=[33817], 99.50th=[36963], 99.90th=[39060], 99.95th=[39584], 00:36:02.378 | 99.99th=[39584] 00:36:02.378 bw ( KiB/s): min= 2672, max= 3216, per=4.14%, avg=2773.60, stdev=133.19, samples=20 00:36:02.378 iops : min= 668, max= 804, avg=693.40, stdev=33.30, samples=20 00:36:02.378 lat (msec) : 10=0.12%, 20=8.75%, 50=91.14% 00:36:02.378 cpu : usr=99.02%, sys=0.69%, ctx=16, majf=0, minf=26 00:36:02.378 IO depths : 1=4.6%, 2=9.5%, 4=21.9%, 8=55.9%, 16=8.0%, 32=0.0%, >=64=0.0% 00:36:02.378 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 complete : 0=0.0%, 4=93.4%, 8=0.9%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 issued rwts: total=6950,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.378 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.378 filename2: (groupid=0, jobs=1): err= 0: pid=353366: Thu Oct 17 17:57:08 2024 00:36:02.378 read: IOPS=697, BW=2792KiB/s (2859kB/s)(27.3MiB/10003msec) 00:36:02.378 slat (usec): min=5, max=118, avg=26.49, stdev=18.96 00:36:02.378 clat (usec): min=8445, max=45656, avg=22714.11, stdev=2713.44 00:36:02.378 lat (usec): min=8452, max=45674, avg=22740.60, stdev=2715.90 00:36:02.378 clat percentiles (usec): 00:36:02.378 | 1.00th=[13042], 5.00th=[16712], 10.00th=[20317], 20.00th=[22414], 00:36:02.378 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.378 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24773], 00:36:02.378 | 99.00th=[31065], 99.50th=[33424], 99.90th=[35390], 99.95th=[45351], 00:36:02.378 | 99.99th=[45876] 00:36:02.378 bw ( KiB/s): min= 2560, max= 3152, per=4.16%, avg=2784.84, stdev=154.62, samples=19 00:36:02.378 iops : min= 640, max= 788, avg=696.21, stdev=38.65, samples=19 00:36:02.378 lat (msec) : 10=0.23%, 20=9.57%, 50=90.20% 00:36:02.378 cpu : usr=98.98%, sys=0.73%, ctx=17, majf=0, minf=33 00:36:02.378 IO depths : 1=4.2%, 2=8.9%, 4=19.6%, 8=58.4%, 16=8.9%, 32=0.0%, >=64=0.0% 00:36:02.378 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 complete : 0=0.0%, 4=92.5%, 8=2.4%, 16=5.1%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 issued rwts: total=6982,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.378 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.378 filename2: (groupid=0, jobs=1): err= 0: pid=353367: Thu Oct 17 17:57:08 2024 00:36:02.378 read: IOPS=697, BW=2790KiB/s (2857kB/s)(27.3MiB/10014msec) 00:36:02.378 slat (usec): min=5, max=124, avg=29.28, stdev=21.24 00:36:02.378 clat (usec): min=11804, max=40616, avg=22691.26, stdev=2478.48 00:36:02.378 lat (usec): min=11820, max=40656, avg=22720.54, stdev=2481.12 00:36:02.378 clat percentiles (usec): 00:36:02.378 | 1.00th=[14222], 5.00th=[16581], 10.00th=[21890], 20.00th=[22414], 00:36:02.378 | 30.00th=[22676], 40.00th=[22938], 50.00th=[22938], 60.00th=[23200], 00:36:02.378 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24511], 00:36:02.378 | 99.00th=[29754], 99.50th=[32900], 99.90th=[38536], 99.95th=[40633], 00:36:02.378 | 99.99th=[40633] 00:36:02.378 bw ( KiB/s): min= 2688, max= 3104, per=4.16%, avg=2788.00, stdev=129.45, samples=20 00:36:02.378 iops : min= 672, max= 776, avg=697.00, stdev=32.36, samples=20 00:36:02.378 lat (msec) : 20=8.46%, 50=91.54% 00:36:02.378 cpu : usr=98.80%, sys=0.91%, ctx=19, majf=0, minf=30 00:36:02.378 IO depths : 1=5.4%, 2=10.8%, 4=22.3%, 8=54.3%, 16=7.3%, 32=0.0%, >=64=0.0% 00:36:02.378 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 complete : 0=0.0%, 4=93.4%, 8=0.9%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 issued rwts: total=6986,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.378 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.378 filename2: (groupid=0, jobs=1): err= 0: pid=353368: Thu Oct 17 17:57:08 2024 00:36:02.378 read: IOPS=679, BW=2716KiB/s (2781kB/s)(26.6MiB/10013msec) 00:36:02.378 slat (usec): min=5, max=146, avg=35.40, stdev=25.43 00:36:02.378 clat (usec): min=12021, max=37097, avg=23245.38, stdev=1351.41 00:36:02.378 lat (usec): min=12047, max=37120, avg=23280.78, stdev=1347.90 00:36:02.378 clat percentiles (usec): 00:36:02.378 | 1.00th=[21627], 5.00th=[22152], 10.00th=[22414], 20.00th=[22414], 00:36:02.378 | 30.00th=[22676], 40.00th=[22938], 50.00th=[23200], 60.00th=[23200], 00:36:02.378 | 70.00th=[23462], 80.00th=[23725], 90.00th=[24249], 95.00th=[24773], 00:36:02.378 | 99.00th=[28181], 99.50th=[31065], 99.90th=[36963], 99.95th=[36963], 00:36:02.378 | 99.99th=[36963] 00:36:02.378 bw ( KiB/s): min= 2432, max= 2816, per=4.05%, avg=2714.95, stdev=109.44, samples=19 00:36:02.378 iops : min= 608, max= 704, avg=678.74, stdev=27.36, samples=19 00:36:02.378 lat (msec) : 20=0.46%, 50=99.54% 00:36:02.378 cpu : usr=99.12%, sys=0.58%, ctx=65, majf=0, minf=35 00:36:02.378 IO depths : 1=6.3%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:36:02.378 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 issued rwts: total=6799,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.378 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.378 filename2: (groupid=0, jobs=1): err= 0: pid=353369: Thu Oct 17 17:57:08 2024 00:36:02.378 read: IOPS=695, BW=2782KiB/s (2849kB/s)(27.2MiB/10003msec) 00:36:02.378 slat (usec): min=5, max=153, avg=25.84, stdev=24.86 00:36:02.378 clat (usec): min=6915, max=45044, avg=22787.35, stdev=2943.75 00:36:02.378 lat (usec): min=6924, max=45061, avg=22813.19, stdev=2944.33 00:36:02.378 clat percentiles (usec): 00:36:02.378 | 1.00th=[13566], 5.00th=[16909], 10.00th=[19530], 20.00th=[22152], 00:36:02.378 | 30.00th=[22414], 40.00th=[22676], 50.00th=[22938], 60.00th=[23200], 00:36:02.378 | 70.00th=[23462], 80.00th=[23987], 90.00th=[24773], 95.00th=[26608], 00:36:02.378 | 99.00th=[31065], 99.50th=[33817], 99.90th=[41681], 99.95th=[41681], 00:36:02.378 | 99.99th=[44827] 00:36:02.378 bw ( KiB/s): min= 2656, max= 2896, per=4.14%, avg=2774.74, stdev=75.67, samples=19 00:36:02.378 iops : min= 664, max= 724, avg=693.68, stdev=18.92, samples=19 00:36:02.378 lat (msec) : 10=0.34%, 20=10.88%, 50=88.78% 00:36:02.378 cpu : usr=98.62%, sys=0.98%, ctx=71, majf=0, minf=28 00:36:02.378 IO depths : 1=2.7%, 2=5.7%, 4=13.3%, 8=66.5%, 16=11.8%, 32=0.0%, >=64=0.0% 00:36:02.378 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 complete : 0=0.0%, 4=91.4%, 8=4.8%, 16=3.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:02.378 issued rwts: total=6958,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:02.378 latency : target=0, window=0, percentile=100.00%, depth=16 00:36:02.378 00:36:02.378 Run status group 0 (all jobs): 00:36:02.378 READ: bw=65.4MiB/s (68.5MB/s), 2716KiB/s-2997KiB/s (2781kB/s-3069kB/s), io=655MiB (687MB), run=10001-10022msec 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@113 -- # destroy_subsystems 0 1 2 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:36:02.378 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 2 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=2 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null2 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # NULL_DIF=1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # bs=8k,16k,128k 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # numjobs=2 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # iodepth=8 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # runtime=5 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # files=1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@117 -- # create_subsystems 0 1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 bdev_null0 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 [2024-10-17 17:57:08.937793] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 bdev_null1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # fio /dev/fd/62 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # create_json_sub_conf 0 1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # config=() 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # local subsystem config 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:02.379 { 00:36:02.379 "params": { 00:36:02.379 "name": "Nvme$subsystem", 00:36:02.379 "trtype": "$TEST_TRANSPORT", 00:36:02.379 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:02.379 "adrfam": "ipv4", 00:36:02.379 "trsvcid": "$NVMF_PORT", 00:36:02.379 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:02.379 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:02.379 "hdgst": ${hdgst:-false}, 00:36:02.379 "ddgst": ${ddgst:-false} 00:36:02.379 }, 00:36:02.379 "method": "bdev_nvme_attach_controller" 00:36:02.379 } 00:36:02.379 EOF 00:36:02.379 )") 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:36:02.379 17:57:08 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:36:02.379 17:57:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:02.379 17:57:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:02.379 { 00:36:02.379 "params": { 00:36:02.379 "name": "Nvme$subsystem", 00:36:02.379 "trtype": "$TEST_TRANSPORT", 00:36:02.379 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:02.380 "adrfam": "ipv4", 00:36:02.380 "trsvcid": "$NVMF_PORT", 00:36:02.380 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:02.380 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:02.380 "hdgst": ${hdgst:-false}, 00:36:02.380 "ddgst": ${ddgst:-false} 00:36:02.380 }, 00:36:02.380 "method": "bdev_nvme_attach_controller" 00:36:02.380 } 00:36:02.380 EOF 00:36:02.380 )") 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # jq . 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@583 -- # IFS=, 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:36:02.380 "params": { 00:36:02.380 "name": "Nvme0", 00:36:02.380 "trtype": "tcp", 00:36:02.380 "traddr": "10.0.0.2", 00:36:02.380 "adrfam": "ipv4", 00:36:02.380 "trsvcid": "4420", 00:36:02.380 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:36:02.380 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:36:02.380 "hdgst": false, 00:36:02.380 "ddgst": false 00:36:02.380 }, 00:36:02.380 "method": "bdev_nvme_attach_controller" 00:36:02.380 },{ 00:36:02.380 "params": { 00:36:02.380 "name": "Nvme1", 00:36:02.380 "trtype": "tcp", 00:36:02.380 "traddr": "10.0.0.2", 00:36:02.380 "adrfam": "ipv4", 00:36:02.380 "trsvcid": "4420", 00:36:02.380 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:36:02.380 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:36:02.380 "hdgst": false, 00:36:02.380 "ddgst": false 00:36:02.380 }, 00:36:02.380 "method": "bdev_nvme_attach_controller" 00:36:02.380 }' 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:36:02.380 17:57:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:36:02.380 filename0: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:36:02.380 ... 00:36:02.380 filename1: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:36:02.380 ... 00:36:02.380 fio-3.35 00:36:02.380 Starting 4 threads 00:36:07.669 00:36:07.669 filename0: (groupid=0, jobs=1): err= 0: pid=356099: Thu Oct 17 17:57:15 2024 00:36:07.669 read: IOPS=2995, BW=23.4MiB/s (24.5MB/s)(117MiB/5002msec) 00:36:07.669 slat (nsec): min=5642, max=58732, avg=9165.86, stdev=3451.11 00:36:07.669 clat (usec): min=978, max=4919, avg=2646.96, stdev=244.88 00:36:07.669 lat (usec): min=997, max=4930, avg=2656.13, stdev=244.63 00:36:07.669 clat percentiles (usec): 00:36:07.669 | 1.00th=[ 1745], 5.00th=[ 2311], 10.00th=[ 2442], 20.00th=[ 2606], 00:36:07.669 | 30.00th=[ 2638], 40.00th=[ 2638], 50.00th=[ 2671], 60.00th=[ 2671], 00:36:07.669 | 70.00th=[ 2671], 80.00th=[ 2704], 90.00th=[ 2835], 95.00th=[ 2933], 00:36:07.669 | 99.00th=[ 3458], 99.50th=[ 3687], 99.90th=[ 4555], 99.95th=[ 4817], 00:36:07.669 | 99.99th=[ 4883] 00:36:07.669 bw ( KiB/s): min=23808, max=24416, per=25.39%, avg=23966.22, stdev=187.35, samples=9 00:36:07.669 iops : min= 2976, max= 3052, avg=2995.78, stdev=23.42, samples=9 00:36:07.669 lat (usec) : 1000=0.01% 00:36:07.669 lat (msec) : 2=1.68%, 4=98.08%, 10=0.23% 00:36:07.669 cpu : usr=97.04%, sys=2.72%, ctx=6, majf=0, minf=88 00:36:07.669 IO depths : 1=0.1%, 2=0.4%, 4=70.0%, 8=29.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:07.669 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.669 complete : 0=0.0%, 4=93.9%, 8=6.1%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.669 issued rwts: total=14983,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:07.669 latency : target=0, window=0, percentile=100.00%, depth=8 00:36:07.669 filename0: (groupid=0, jobs=1): err= 0: pid=356101: Thu Oct 17 17:57:15 2024 00:36:07.669 read: IOPS=2959, BW=23.1MiB/s (24.2MB/s)(117MiB/5042msec) 00:36:07.669 slat (nsec): min=5637, max=53060, avg=7240.93, stdev=2874.30 00:36:07.669 clat (usec): min=871, max=41703, avg=2674.43, stdev=750.40 00:36:07.669 lat (usec): min=878, max=41710, avg=2681.67, stdev=750.41 00:36:07.669 clat percentiles (usec): 00:36:07.669 | 1.00th=[ 2024], 5.00th=[ 2311], 10.00th=[ 2474], 20.00th=[ 2606], 00:36:07.669 | 30.00th=[ 2638], 40.00th=[ 2638], 50.00th=[ 2671], 60.00th=[ 2671], 00:36:07.669 | 70.00th=[ 2671], 80.00th=[ 2704], 90.00th=[ 2835], 95.00th=[ 2933], 00:36:07.669 | 99.00th=[ 3720], 99.50th=[ 3982], 99.90th=[ 4621], 99.95th=[ 4817], 00:36:07.669 | 99.99th=[41681] 00:36:07.669 bw ( KiB/s): min=23615, max=24112, per=25.29%, avg=23873.50, stdev=141.40, samples=10 00:36:07.669 iops : min= 2951, max= 3014, avg=2984.10, stdev=17.85, samples=10 00:36:07.669 lat (usec) : 1000=0.02% 00:36:07.669 lat (msec) : 2=0.84%, 4=98.79%, 10=0.32%, 50=0.03% 00:36:07.669 cpu : usr=96.67%, sys=3.07%, ctx=7, majf=0, minf=116 00:36:07.669 IO depths : 1=0.1%, 2=0.2%, 4=72.5%, 8=27.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:07.669 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.669 complete : 0=0.0%, 4=92.1%, 8=7.9%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.669 issued rwts: total=14923,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:07.669 latency : target=0, window=0, percentile=100.00%, depth=8 00:36:07.669 filename1: (groupid=0, jobs=1): err= 0: pid=356103: Thu Oct 17 17:57:15 2024 00:36:07.669 read: IOPS=2914, BW=22.8MiB/s (23.9MB/s)(114MiB/5001msec) 00:36:07.669 slat (nsec): min=5651, max=73733, avg=8298.81, stdev=3473.16 00:36:07.669 clat (usec): min=980, max=5155, avg=2722.28, stdev=310.92 00:36:07.669 lat (usec): min=988, max=5165, avg=2730.57, stdev=311.13 00:36:07.669 clat percentiles (usec): 00:36:07.669 | 1.00th=[ 2073], 5.00th=[ 2442], 10.00th=[ 2540], 20.00th=[ 2638], 00:36:07.669 | 30.00th=[ 2638], 40.00th=[ 2638], 50.00th=[ 2671], 60.00th=[ 2671], 00:36:07.669 | 70.00th=[ 2704], 80.00th=[ 2737], 90.00th=[ 2966], 95.00th=[ 3326], 00:36:07.669 | 99.00th=[ 4047], 99.50th=[ 4359], 99.90th=[ 4948], 99.95th=[ 5014], 00:36:07.669 | 99.99th=[ 5145] 00:36:07.669 bw ( KiB/s): min=23024, max=23648, per=24.73%, avg=23340.44, stdev=243.81, samples=9 00:36:07.669 iops : min= 2878, max= 2956, avg=2917.56, stdev=30.48, samples=9 00:36:07.669 lat (usec) : 1000=0.02% 00:36:07.669 lat (msec) : 2=0.79%, 4=98.08%, 10=1.11% 00:36:07.669 cpu : usr=96.74%, sys=3.02%, ctx=7, majf=0, minf=76 00:36:07.669 IO depths : 1=0.1%, 2=0.5%, 4=72.1%, 8=27.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:07.669 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.669 complete : 0=0.0%, 4=92.2%, 8=7.8%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.669 issued rwts: total=14575,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:07.669 latency : target=0, window=0, percentile=100.00%, depth=8 00:36:07.669 filename1: (groupid=0, jobs=1): err= 0: pid=356104: Thu Oct 17 17:57:15 2024 00:36:07.669 read: IOPS=3000, BW=23.4MiB/s (24.6MB/s)(117MiB/5001msec) 00:36:07.669 slat (nsec): min=5635, max=56199, avg=6998.85, stdev=3025.08 00:36:07.669 clat (usec): min=1010, max=4632, avg=2647.12, stdev=253.01 00:36:07.669 lat (usec): min=1016, max=4638, avg=2654.12, stdev=253.08 00:36:07.669 clat percentiles (usec): 00:36:07.669 | 1.00th=[ 1860], 5.00th=[ 2245], 10.00th=[ 2409], 20.00th=[ 2606], 00:36:07.669 | 30.00th=[ 2638], 40.00th=[ 2638], 50.00th=[ 2671], 60.00th=[ 2671], 00:36:07.669 | 70.00th=[ 2671], 80.00th=[ 2704], 90.00th=[ 2835], 95.00th=[ 2933], 00:36:07.669 | 99.00th=[ 3589], 99.50th=[ 3851], 99.90th=[ 4293], 99.95th=[ 4555], 00:36:07.669 | 99.99th=[ 4621] 00:36:07.669 bw ( KiB/s): min=23872, max=24272, per=25.44%, avg=24008.89, stdev=155.15, samples=9 00:36:07.670 iops : min= 2984, max= 3034, avg=3001.11, stdev=19.39, samples=9 00:36:07.670 lat (msec) : 2=1.95%, 4=97.75%, 10=0.30% 00:36:07.670 cpu : usr=96.44%, sys=3.28%, ctx=17, majf=0, minf=89 00:36:07.670 IO depths : 1=0.1%, 2=0.3%, 4=70.8%, 8=28.8%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:07.670 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.670 complete : 0=0.0%, 4=93.4%, 8=6.6%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:07.670 issued rwts: total=15008,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:07.670 latency : target=0, window=0, percentile=100.00%, depth=8 00:36:07.670 00:36:07.670 Run status group 0 (all jobs): 00:36:07.670 READ: bw=92.2MiB/s (96.7MB/s), 22.8MiB/s-23.4MiB/s (23.9MB/s-24.6MB/s), io=465MiB (487MB), run=5001-5042msec 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@119 -- # destroy_subsystems 0 1 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 00:36:07.670 real 0m24.425s 00:36:07.670 user 5m12.295s 00:36:07.670 sys 0m4.625s 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1126 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 ************************************ 00:36:07.670 END TEST fio_dif_rand_params 00:36:07.670 ************************************ 00:36:07.670 17:57:15 nvmf_dif -- target/dif.sh@144 -- # run_test fio_dif_digest fio_dif_digest 00:36:07.670 17:57:15 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:36:07.670 17:57:15 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 ************************************ 00:36:07.670 START TEST fio_dif_digest 00:36:07.670 ************************************ 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1125 -- # fio_dif_digest 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@123 -- # local NULL_DIF 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@124 -- # local bs numjobs runtime iodepth files 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@125 -- # local hdgst ddgst 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # NULL_DIF=3 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # bs=128k,128k,128k 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # numjobs=3 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # iodepth=3 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # runtime=10 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # hdgst=true 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # ddgst=true 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@130 -- # create_subsystems 0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@28 -- # local sub 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@30 -- # for sub in "$@" 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@31 -- # create_subsystem 0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@18 -- # local sub_id=0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 bdev_null0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:36:07.670 [2024-10-17 17:57:15.492898] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # fio /dev/fd/62 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # create_json_sub_conf 0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@558 -- # config=() 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@558 -- # local subsystem config 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:07.670 { 00:36:07.670 "params": { 00:36:07.670 "name": "Nvme$subsystem", 00:36:07.670 "trtype": "$TEST_TRANSPORT", 00:36:07.670 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:07.670 "adrfam": "ipv4", 00:36:07.670 "trsvcid": "$NVMF_PORT", 00:36:07.670 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:07.670 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:07.670 "hdgst": ${hdgst:-false}, 00:36:07.670 "ddgst": ${ddgst:-false} 00:36:07.670 }, 00:36:07.670 "method": "bdev_nvme_attach_controller" 00:36:07.670 } 00:36:07.670 EOF 00:36:07.670 )") 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # gen_fio_conf 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@54 -- # local file 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1339 -- # local sanitizers 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@56 -- # cat 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1341 -- # shift 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1343 -- # local asan_lib= 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@580 -- # cat 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file = 1 )) 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # grep libasan 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file <= files )) 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@582 -- # jq . 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@583 -- # IFS=, 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:36:07.670 "params": { 00:36:07.670 "name": "Nvme0", 00:36:07.670 "trtype": "tcp", 00:36:07.670 "traddr": "10.0.0.2", 00:36:07.670 "adrfam": "ipv4", 00:36:07.670 "trsvcid": "4420", 00:36:07.670 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:36:07.670 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:36:07.670 "hdgst": true, 00:36:07.670 "ddgst": true 00:36:07.670 }, 00:36:07.670 "method": "bdev_nvme_attach_controller" 00:36:07.670 }' 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # asan_lib= 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # asan_lib= 00:36:07.670 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:36:07.671 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:36:07.671 17:57:15 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:36:08.273 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:36:08.273 ... 00:36:08.273 fio-3.35 00:36:08.273 Starting 3 threads 00:36:20.513 00:36:20.513 filename0: (groupid=0, jobs=1): err= 0: pid=357635: Thu Oct 17 17:57:26 2024 00:36:20.513 read: IOPS=289, BW=36.1MiB/s (37.9MB/s)(363MiB/10047msec) 00:36:20.513 slat (nsec): min=6008, max=64882, avg=8895.01, stdev=2293.68 00:36:20.513 clat (usec): min=4612, max=92306, avg=10353.83, stdev=8207.05 00:36:20.513 lat (usec): min=4620, max=92314, avg=10362.73, stdev=8207.06 00:36:20.513 clat percentiles (usec): 00:36:20.513 | 1.00th=[ 7111], 5.00th=[ 7701], 10.00th=[ 7963], 20.00th=[ 8225], 00:36:20.513 | 30.00th=[ 8455], 40.00th=[ 8586], 50.00th=[ 8848], 60.00th=[ 8979], 00:36:20.513 | 70.00th=[ 9241], 80.00th=[ 9503], 90.00th=[ 9896], 95.00th=[10814], 00:36:20.513 | 99.00th=[51119], 99.50th=[51643], 99.90th=[90702], 99.95th=[92799], 00:36:20.513 | 99.99th=[92799] 00:36:20.513 bw ( KiB/s): min=24576, max=44288, per=32.31%, avg=37145.60, stdev=5623.36, samples=20 00:36:20.513 iops : min= 192, max= 346, avg=290.20, stdev=43.93, samples=20 00:36:20.513 lat (msec) : 10=90.56%, 20=5.92%, 50=1.31%, 100=2.20% 00:36:20.513 cpu : usr=94.00%, sys=5.75%, ctx=27, majf=0, minf=147 00:36:20.513 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:20.513 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:20.513 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:20.513 issued rwts: total=2904,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:20.513 latency : target=0, window=0, percentile=100.00%, depth=3 00:36:20.513 filename0: (groupid=0, jobs=1): err= 0: pid=357636: Thu Oct 17 17:57:26 2024 00:36:20.513 read: IOPS=305, BW=38.2MiB/s (40.1MB/s)(384MiB/10045msec) 00:36:20.513 slat (nsec): min=5973, max=57333, avg=9111.77, stdev=1930.76 00:36:20.513 clat (usec): min=5772, max=53650, avg=9784.39, stdev=2108.88 00:36:20.513 lat (usec): min=5782, max=53686, avg=9793.51, stdev=2109.19 00:36:20.513 clat percentiles (usec): 00:36:20.513 | 1.00th=[ 6783], 5.00th=[ 7308], 10.00th=[ 7635], 20.00th=[ 8717], 00:36:20.513 | 30.00th=[ 9372], 40.00th=[ 9634], 50.00th=[ 9896], 60.00th=[10159], 00:36:20.513 | 70.00th=[10421], 80.00th=[10814], 90.00th=[11207], 95.00th=[11469], 00:36:20.513 | 99.00th=[12256], 99.50th=[12649], 99.90th=[50594], 99.95th=[52691], 00:36:20.513 | 99.99th=[53740] 00:36:20.513 bw ( KiB/s): min=34560, max=43264, per=34.18%, avg=39296.00, stdev=2102.02, samples=20 00:36:20.513 iops : min= 270, max= 338, avg=307.00, stdev=16.42, samples=20 00:36:20.513 lat (msec) : 10=53.19%, 20=46.65%, 50=0.03%, 100=0.13% 00:36:20.513 cpu : usr=95.74%, sys=4.02%, ctx=16, majf=0, minf=200 00:36:20.513 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:20.513 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:20.513 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:20.513 issued rwts: total=3072,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:20.513 latency : target=0, window=0, percentile=100.00%, depth=3 00:36:20.513 filename0: (groupid=0, jobs=1): err= 0: pid=357637: Thu Oct 17 17:57:26 2024 00:36:20.513 read: IOPS=303, BW=37.9MiB/s (39.8MB/s)(381MiB/10046msec) 00:36:20.513 slat (nsec): min=5965, max=30705, avg=9017.09, stdev=1762.18 00:36:20.513 clat (usec): min=4939, max=52400, avg=9862.96, stdev=2114.04 00:36:20.513 lat (usec): min=4948, max=52431, avg=9871.98, stdev=2114.22 00:36:20.513 clat percentiles (usec): 00:36:20.513 | 1.00th=[ 6587], 5.00th=[ 7242], 10.00th=[ 7701], 20.00th=[ 8717], 00:36:20.513 | 30.00th=[ 9372], 40.00th=[ 9765], 50.00th=[10028], 60.00th=[10290], 00:36:20.513 | 70.00th=[10552], 80.00th=[10814], 90.00th=[11338], 95.00th=[11731], 00:36:20.513 | 99.00th=[12518], 99.50th=[12649], 99.90th=[49021], 99.95th=[52167], 00:36:20.513 | 99.99th=[52167] 00:36:20.513 bw ( KiB/s): min=36023, max=44288, per=33.91%, avg=38985.15, stdev=2067.05, samples=20 00:36:20.513 iops : min= 281, max= 346, avg=304.55, stdev=16.18, samples=20 00:36:20.513 lat (msec) : 10=48.29%, 20=51.54%, 50=0.07%, 100=0.10% 00:36:20.513 cpu : usr=95.70%, sys=4.04%, ctx=19, majf=0, minf=156 00:36:20.513 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:20.513 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:20.513 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:20.513 issued rwts: total=3048,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:20.513 latency : target=0, window=0, percentile=100.00%, depth=3 00:36:20.513 00:36:20.513 Run status group 0 (all jobs): 00:36:20.513 READ: bw=112MiB/s (118MB/s), 36.1MiB/s-38.2MiB/s (37.9MB/s-40.1MB/s), io=1128MiB (1183MB), run=10045-10047msec 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- target/dif.sh@132 -- # destroy_subsystems 0 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- target/dif.sh@43 -- # local sub 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- target/dif.sh@45 -- # for sub in "$@" 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- target/dif.sh@46 -- # destroy_subsystem 0 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- target/dif.sh@36 -- # local sub_id=0 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:20.513 00:36:20.513 real 0m11.262s 00:36:20.513 user 0m43.100s 00:36:20.513 sys 0m1.731s 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1126 -- # xtrace_disable 00:36:20.513 17:57:26 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:36:20.513 ************************************ 00:36:20.513 END TEST fio_dif_digest 00:36:20.513 ************************************ 00:36:20.513 17:57:26 nvmf_dif -- target/dif.sh@146 -- # trap - SIGINT SIGTERM EXIT 00:36:20.513 17:57:26 nvmf_dif -- target/dif.sh@147 -- # nvmftestfini 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@514 -- # nvmfcleanup 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@121 -- # sync 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@124 -- # set +e 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:20.513 rmmod nvme_tcp 00:36:20.513 rmmod nvme_fabrics 00:36:20.513 rmmod nvme_keyring 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@128 -- # set -e 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@129 -- # return 0 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@515 -- # '[' -n 346911 ']' 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@516 -- # killprocess 346911 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@950 -- # '[' -z 346911 ']' 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@954 -- # kill -0 346911 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@955 -- # uname 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 346911 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@968 -- # echo 'killing process with pid 346911' 00:36:20.513 killing process with pid 346911 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@969 -- # kill 346911 00:36:20.513 17:57:26 nvmf_dif -- common/autotest_common.sh@974 -- # wait 346911 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@518 -- # '[' iso == iso ']' 00:36:20.513 17:57:26 nvmf_dif -- nvmf/common.sh@519 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:36:23.063 Waiting for block devices as requested 00:36:23.063 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:36:23.063 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:36:23.063 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:36:23.063 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:36:23.063 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:36:23.063 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:36:23.063 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:36:23.324 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:36:23.324 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:36:23.585 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:36:23.586 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:36:23.586 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:36:23.847 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:36:23.847 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:36:23.847 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:36:24.108 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:36:24.108 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@297 -- # iptr 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@789 -- # iptables-save 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@789 -- # iptables-restore 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:24.369 17:57:32 nvmf_dif -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:24.369 17:57:32 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:36:24.369 17:57:32 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:26.918 17:57:34 nvmf_dif -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:26.918 00:36:26.918 real 1m18.809s 00:36:26.918 user 7m49.121s 00:36:26.918 sys 0m22.079s 00:36:26.918 17:57:34 nvmf_dif -- common/autotest_common.sh@1126 -- # xtrace_disable 00:36:26.918 17:57:34 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:36:26.918 ************************************ 00:36:26.918 END TEST nvmf_dif 00:36:26.918 ************************************ 00:36:26.918 17:57:34 -- spdk/autotest.sh@286 -- # run_test nvmf_abort_qd_sizes /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:36:26.918 17:57:34 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:36:26.918 17:57:34 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:36:26.918 17:57:34 -- common/autotest_common.sh@10 -- # set +x 00:36:26.918 ************************************ 00:36:26.918 START TEST nvmf_abort_qd_sizes 00:36:26.918 ************************************ 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:36:26.918 * Looking for test storage... 00:36:26.918 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1691 -- # lcov --version 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # IFS=.-: 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # read -ra ver1 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # IFS=.-: 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # read -ra ver2 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@338 -- # local 'op=<' 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@340 -- # ver1_l=2 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@341 -- # ver2_l=1 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@344 -- # case "$op" in 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@345 -- # : 1 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:26.918 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # decimal 1 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=1 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 1 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # ver1[v]=1 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # decimal 2 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=2 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 2 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # ver2[v]=2 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # return 0 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:36:26.919 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:26.919 --rc genhtml_branch_coverage=1 00:36:26.919 --rc genhtml_function_coverage=1 00:36:26.919 --rc genhtml_legend=1 00:36:26.919 --rc geninfo_all_blocks=1 00:36:26.919 --rc geninfo_unexecuted_blocks=1 00:36:26.919 00:36:26.919 ' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:36:26.919 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:26.919 --rc genhtml_branch_coverage=1 00:36:26.919 --rc genhtml_function_coverage=1 00:36:26.919 --rc genhtml_legend=1 00:36:26.919 --rc geninfo_all_blocks=1 00:36:26.919 --rc geninfo_unexecuted_blocks=1 00:36:26.919 00:36:26.919 ' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:36:26.919 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:26.919 --rc genhtml_branch_coverage=1 00:36:26.919 --rc genhtml_function_coverage=1 00:36:26.919 --rc genhtml_legend=1 00:36:26.919 --rc geninfo_all_blocks=1 00:36:26.919 --rc geninfo_unexecuted_blocks=1 00:36:26.919 00:36:26.919 ' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:36:26.919 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:26.919 --rc genhtml_branch_coverage=1 00:36:26.919 --rc genhtml_function_coverage=1 00:36:26.919 --rc genhtml_legend=1 00:36:26.919 --rc geninfo_all_blocks=1 00:36:26.919 --rc geninfo_unexecuted_blocks=1 00:36:26.919 00:36:26.919 ' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # uname -s 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@15 -- # shopt -s extglob 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- paths/export.sh@5 -- # export PATH 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@51 -- # : 0 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:36:26.919 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@70 -- # nvmftestinit 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@474 -- # prepare_net_devs 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@436 -- # local -g is_hw=no 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@438 -- # remove_spdk_ns 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- nvmf/common.sh@309 -- # xtrace_disable 00:36:26.919 17:57:34 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # pci_devs=() 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # net_devs=() 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # e810=() 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # local -ga e810 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # x722=() 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # local -ga x722 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # mlx=() 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # local -ga mlx 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:35.070 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.0 (0x8086 - 0x159b)' 00:36:35.071 Found 0000:31:00.0 (0x8086 - 0x159b) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:31:00.1 (0x8086 - 0x159b)' 00:36:35.071 Found 0000:31:00.1 (0x8086 - 0x159b) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.0: cvl_0_0' 00:36:35.071 Found net devices under 0000:31:00.0: cvl_0_0 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:31:00.1: cvl_0_1' 00:36:35.071 Found net devices under 0000:31:00.1: cvl_0_1 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # is_hw=yes 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:35.071 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:35.071 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.589 ms 00:36:35.071 00:36:35.071 --- 10.0.0.2 ping statistics --- 00:36:35.071 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:35.071 rtt min/avg/max/mdev = 0.589/0.589/0.589/0.000 ms 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:35.071 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:35.071 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.188 ms 00:36:35.071 00:36:35.071 --- 10.0.0.1 ping statistics --- 00:36:35.071 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:35.071 rtt min/avg/max/mdev = 0.188/0.188/0.188/0.000 ms 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@448 -- # return 0 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@476 -- # '[' iso == iso ']' 00:36:35.071 17:57:41 nvmf_abort_qd_sizes -- nvmf/common.sh@477 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:36:37.618 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:36:37.618 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:36:38.191 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@71 -- # nvmfappstart -m 0xf 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@724 -- # xtrace_disable 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@507 -- # nvmfpid=367121 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@508 -- # waitforlisten 367121 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xf 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@831 -- # '[' -z 367121 ']' 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@836 -- # local max_retries=100 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:38.192 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@840 -- # xtrace_disable 00:36:38.192 17:57:45 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:36:38.192 [2024-10-17 17:57:45.957991] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:36:38.192 [2024-10-17 17:57:45.958039] [ DPDK EAL parameters: nvmf -c 0xf --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:38.192 [2024-10-17 17:57:46.043397] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:36:38.192 [2024-10-17 17:57:46.097445] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:38.192 [2024-10-17 17:57:46.097484] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:38.192 [2024-10-17 17:57:46.097492] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:38.192 [2024-10-17 17:57:46.097499] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:38.192 [2024-10-17 17:57:46.097505] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:38.192 [2024-10-17 17:57:46.099159] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:38.192 [2024-10-17 17:57:46.099310] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:36:38.192 [2024-10-17 17:57:46.099448] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:38.192 [2024-10-17 17:57:46.099449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- common/autotest_common.sh@864 -- # return 0 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- common/autotest_common.sh@730 -- # xtrace_disable 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@73 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini || :; clean_kernel_target' SIGINT SIGTERM EXIT 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # mapfile -t nvmes 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # nvme_in_userspace 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@312 -- # local bdf bdfs 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@313 -- # local nvmes 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@315 -- # [[ -n 0000:65:00.0 ]] 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@316 -- # nvmes=(${pci_bus_cache["0x010802"]}) 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@321 -- # for bdf in "${nvmes[@]}" 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@322 -- # [[ -e /sys/bus/pci/drivers/nvme/0000:65:00.0 ]] 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # uname -s 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # [[ Linux == FreeBSD ]] 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@326 -- # bdfs+=("$bdf") 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@328 -- # (( 1 )) 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- scripts/common.sh@329 -- # printf '%s\n' 0000:65:00.0 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@76 -- # (( 1 > 0 )) 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@78 -- # nvme=0000:65:00.0 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@80 -- # run_test spdk_target_abort spdk_target 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- common/autotest_common.sh@1107 -- # xtrace_disable 00:36:39.135 17:57:46 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:36:39.135 ************************************ 00:36:39.135 START TEST spdk_target_abort 00:36:39.135 ************************************ 00:36:39.135 17:57:46 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1125 -- # spdk_target 00:36:39.135 17:57:46 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@43 -- # local name=spdk_target 00:36:39.135 17:57:46 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@45 -- # rpc_cmd bdev_nvme_attach_controller -t pcie -a 0000:65:00.0 -b spdk_target 00:36:39.135 17:57:46 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:39.135 17:57:46 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:39.396 spdk_targetn1 00:36:39.396 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:39.396 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@47 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:36:39.396 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:39.396 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:39.396 [2024-10-17 17:57:47.168565] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:39.396 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:testnqn -a -s SPDKISFASTANDAWESOME 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@49 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:testnqn spdk_targetn1 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@50 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:testnqn -t tcp -a 10.0.0.2 -s 4420 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:39.397 [2024-10-17 17:57:47.229891] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@52 -- # rabort tcp IPv4 10.0.0.2 4420 nqn.2016-06.io.spdk:testnqn 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.2 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2' 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:36:39.397 17:57:47 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:36:39.658 [2024-10-17 17:57:47.382395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:32 len:8 PRP1 0x200004ac0000 PRP2 0x0 00:36:39.658 [2024-10-17 17:57:47.382428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:0006 p:1 m:0 dnr:0 00:36:39.658 [2024-10-17 17:57:47.390228] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:191 nsid:1 lba:384 len:8 PRP1 0x200004ac6000 PRP2 0x0 00:36:39.658 [2024-10-17 17:57:47.390249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:191 cdw0:0 sqhd:0031 p:1 m:0 dnr:0 00:36:39.658 [2024-10-17 17:57:47.422186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:1704 len:8 PRP1 0x200004ac6000 PRP2 0x0 00:36:39.658 [2024-10-17 17:57:47.422208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:00d6 p:1 m:0 dnr:0 00:36:39.658 [2024-10-17 17:57:47.438270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:2384 len:8 PRP1 0x200004ac0000 PRP2 0x0 00:36:39.658 [2024-10-17 17:57:47.438291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:42.961 Initializing NVMe Controllers 00:36:42.961 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:36:42.961 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:36:42.961 Initialization complete. Launching workers. 00:36:42.961 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 15628, failed: 4 00:36:42.961 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2276, failed to submit 13356 00:36:42.961 success 660, unsuccessful 1616, failed 0 00:36:42.961 17:57:50 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:36:42.961 17:57:50 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:36:42.961 [2024-10-17 17:57:50.671977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:187 nsid:1 lba:432 len:8 PRP1 0x200004e4a000 PRP2 0x0 00:36:42.961 [2024-10-17 17:57:50.672019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:187 cdw0:0 sqhd:0048 p:1 m:0 dnr:0 00:36:42.961 [2024-10-17 17:57:50.719894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:178 nsid:1 lba:1456 len:8 PRP1 0x200004e56000 PRP2 0x0 00:36:42.961 [2024-10-17 17:57:50.719920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:178 cdw0:0 sqhd:00bb p:1 m:0 dnr:0 00:36:42.961 [2024-10-17 17:57:50.743757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:189 nsid:1 lba:2000 len:8 PRP1 0x200004e40000 PRP2 0x0 00:36:42.961 [2024-10-17 17:57:50.743779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:189 cdw0:0 sqhd:00fb p:1 m:0 dnr:0 00:36:43.903 [2024-10-17 17:57:51.469935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:179 nsid:1 lba:18440 len:8 PRP1 0x200004e44000 PRP2 0x0 00:36:43.903 [2024-10-17 17:57:51.469970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:179 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:36:44.164 [2024-10-17 17:57:51.860023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:187 nsid:1 lba:27048 len:8 PRP1 0x200004e58000 PRP2 0x0 00:36:44.164 [2024-10-17 17:57:51.860052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:187 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:36:46.077 Initializing NVMe Controllers 00:36:46.077 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:36:46.077 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:36:46.077 Initialization complete. Launching workers. 00:36:46.077 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 8414, failed: 5 00:36:46.077 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1223, failed to submit 7196 00:36:46.077 success 345, unsuccessful 878, failed 0 00:36:46.077 17:57:53 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:36:46.077 17:57:53 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:36:46.077 [2024-10-17 17:57:53.933040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:148 nsid:1 lba:2784 len:8 PRP1 0x200004b00000 PRP2 0x0 00:36:46.077 [2024-10-17 17:57:53.933067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:148 cdw0:0 sqhd:0077 p:1 m:0 dnr:0 00:36:47.460 [2024-10-17 17:57:55.300339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:161104 len:8 PRP1 0x200004af8000 PRP2 0x0 00:36:47.460 [2024-10-17 17:57:55.300369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:00bf p:0 m:0 dnr:0 00:36:49.499 Initializing NVMe Controllers 00:36:49.499 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:36:49.499 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:36:49.499 Initialization complete. Launching workers. 00:36:49.499 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 43761, failed: 2 00:36:49.499 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2782, failed to submit 40981 00:36:49.499 success 591, unsuccessful 2191, failed 0 00:36:49.499 17:57:56 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@54 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:testnqn 00:36:49.499 17:57:56 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:49.499 17:57:56 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:49.499 17:57:56 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:49.499 17:57:56 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@55 -- # rpc_cmd bdev_nvme_detach_controller spdk_target 00:36:49.499 17:57:56 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:49.499 17:57:56 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@61 -- # killprocess 367121 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@950 -- # '[' -z 367121 ']' 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@954 -- # kill -0 367121 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@955 -- # uname 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 367121 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@968 -- # echo 'killing process with pid 367121' 00:36:51.410 killing process with pid 367121 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@969 -- # kill 367121 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@974 -- # wait 367121 00:36:51.410 00:36:51.410 real 0m12.118s 00:36:51.410 user 0m49.349s 00:36:51.410 sys 0m2.009s 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:36:51.410 17:57:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:36:51.410 ************************************ 00:36:51.410 END TEST spdk_target_abort 00:36:51.410 ************************************ 00:36:51.410 17:57:59 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@81 -- # run_test kernel_target_abort kernel_target 00:36:51.410 17:57:59 nvmf_abort_qd_sizes -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:36:51.410 17:57:59 nvmf_abort_qd_sizes -- common/autotest_common.sh@1107 -- # xtrace_disable 00:36:51.410 17:57:59 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:36:51.410 ************************************ 00:36:51.410 START TEST kernel_target_abort 00:36:51.410 ************************************ 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1125 -- # kernel_target 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # get_main_ns_ip 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@767 -- # local ip 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@768 -- # ip_candidates=() 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@768 -- # local -A ip_candidates 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@658 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@660 -- # nvmet=/sys/kernel/config/nvmet 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@661 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@662 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@663 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@665 -- # local block nvme 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@667 -- # [[ ! -e /sys/module/nvmet ]] 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@668 -- # modprobe nvmet 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@671 -- # [[ -e /sys/kernel/config/nvmet ]] 00:36:51.410 17:57:59 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@673 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:36:54.713 Waiting for block devices as requested 00:36:54.713 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:36:54.974 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:36:54.974 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:36:54.974 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:36:54.974 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:36:55.236 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:36:55.236 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:36:55.236 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:36:55.497 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:36:55.758 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:36:55.758 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:36:55.758 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:36:56.019 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:36:56.019 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:36:56.019 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:36:56.019 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:36:56.280 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@676 -- # for block in /sys/block/nvme* 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@677 -- # [[ -e /sys/block/nvme0n1 ]] 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@678 -- # is_block_zoned nvme0n1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@679 -- # block_in_use nvme0n1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:36:56.541 No valid GPT data, bailing 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # pt= 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@395 -- # return 1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@679 -- # nvme=/dev/nvme0n1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@682 -- # [[ -b /dev/nvme0n1 ]] 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@684 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@685 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@691 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@693 -- # echo 1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@694 -- # echo /dev/nvme0n1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@695 -- # echo 1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@697 -- # echo 10.0.0.1 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@698 -- # echo tcp 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@699 -- # echo 4420 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@700 -- # echo ipv4 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@703 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:36:56.541 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@706 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 --hostid=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 -a 10.0.0.1 -t tcp -s 4420 00:36:56.803 00:36:56.803 Discovery Log Number of Records 2, Generation counter 2 00:36:56.803 =====Discovery Log Entry 0====== 00:36:56.803 trtype: tcp 00:36:56.803 adrfam: ipv4 00:36:56.803 subtype: current discovery subsystem 00:36:56.803 treq: not specified, sq flow control disable supported 00:36:56.803 portid: 1 00:36:56.803 trsvcid: 4420 00:36:56.803 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:36:56.803 traddr: 10.0.0.1 00:36:56.803 eflags: none 00:36:56.803 sectype: none 00:36:56.803 =====Discovery Log Entry 1====== 00:36:56.803 trtype: tcp 00:36:56.803 adrfam: ipv4 00:36:56.803 subtype: nvme subsystem 00:36:56.803 treq: not specified, sq flow control disable supported 00:36:56.803 portid: 1 00:36:56.803 trsvcid: 4420 00:36:56.803 subnqn: nqn.2016-06.io.spdk:testnqn 00:36:56.803 traddr: 10.0.0.1 00:36:56.803 eflags: none 00:36:56.803 sectype: none 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@66 -- # rabort tcp IPv4 10.0.0.1 4420 nqn.2016-06.io.spdk:testnqn 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.1 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1' 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420' 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:36:56.803 17:58:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:37:00.108 Initializing NVMe Controllers 00:37:00.108 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:37:00.108 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:37:00.108 Initialization complete. Launching workers. 00:37:00.108 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 67161, failed: 0 00:37:00.108 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 67161, failed to submit 0 00:37:00.108 success 0, unsuccessful 67161, failed 0 00:37:00.108 17:58:07 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:37:00.108 17:58:07 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:37:03.411 Initializing NVMe Controllers 00:37:03.411 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:37:03.411 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:37:03.411 Initialization complete. Launching workers. 00:37:03.411 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 116514, failed: 0 00:37:03.412 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 29306, failed to submit 87208 00:37:03.412 success 0, unsuccessful 29306, failed 0 00:37:03.412 17:58:10 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:37:03.412 17:58:10 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:37:05.961 Initializing NVMe Controllers 00:37:05.961 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:37:05.961 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:37:05.961 Initialization complete. Launching workers. 00:37:05.961 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 147008, failed: 0 00:37:05.961 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 36810, failed to submit 110198 00:37:05.961 success 0, unsuccessful 36810, failed 0 00:37:05.961 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@67 -- # clean_kernel_target 00:37:05.961 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@710 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:37:05.961 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@712 -- # echo 0 00:37:05.961 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@714 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:37:05.961 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@715 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:37:05.961 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@716 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:37:06.222 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:37:06.222 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@719 -- # modules=(/sys/module/nvmet/holders/*) 00:37:06.222 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@721 -- # modprobe -r nvmet_tcp nvmet 00:37:06.222 17:58:13 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@724 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:37:09.527 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:37:09.527 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:37:09.788 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:37:09.788 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:37:09.788 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:37:11.704 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:37:11.704 00:37:11.704 real 0m20.551s 00:37:11.704 user 0m9.797s 00:37:11.704 sys 0m6.332s 00:37:11.704 17:58:19 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:37:11.704 17:58:19 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@10 -- # set +x 00:37:11.704 ************************************ 00:37:11.704 END TEST kernel_target_abort 00:37:11.704 ************************************ 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@84 -- # nvmftestfini 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@514 -- # nvmfcleanup 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@121 -- # sync 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@124 -- # set +e 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@125 -- # for i in {1..20} 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:37:11.966 rmmod nvme_tcp 00:37:11.966 rmmod nvme_fabrics 00:37:11.966 rmmod nvme_keyring 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@128 -- # set -e 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@129 -- # return 0 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@515 -- # '[' -n 367121 ']' 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@516 -- # killprocess 367121 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- common/autotest_common.sh@950 -- # '[' -z 367121 ']' 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- common/autotest_common.sh@954 -- # kill -0 367121 00:37:11.966 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (367121) - No such process 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- common/autotest_common.sh@977 -- # echo 'Process with pid 367121 is not found' 00:37:11.966 Process with pid 367121 is not found 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@518 -- # '[' iso == iso ']' 00:37:11.966 17:58:19 nvmf_abort_qd_sizes -- nvmf/common.sh@519 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:37:15.274 Waiting for block devices as requested 00:37:15.536 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:37:15.536 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:37:15.536 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:37:15.536 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:37:15.797 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:37:15.798 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:37:15.798 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:37:16.059 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:37:16.059 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:37:16.320 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:37:16.320 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:37:16.320 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:37:16.582 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:37:16.582 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:37:16.582 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:37:16.842 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:37:16.842 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@297 -- # iptr 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@789 -- # iptables-save 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@789 -- # iptables-restore 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@302 -- # remove_spdk_ns 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:37:17.104 17:58:24 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:19.651 17:58:27 nvmf_abort_qd_sizes -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:37:19.651 00:37:19.651 real 0m52.725s 00:37:19.651 user 1m4.585s 00:37:19.651 sys 0m19.521s 00:37:19.651 17:58:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@1126 -- # xtrace_disable 00:37:19.651 17:58:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:37:19.651 ************************************ 00:37:19.651 END TEST nvmf_abort_qd_sizes 00:37:19.651 ************************************ 00:37:19.651 17:58:27 -- spdk/autotest.sh@288 -- # run_test keyring_file /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:37:19.651 17:58:27 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:37:19.651 17:58:27 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:37:19.651 17:58:27 -- common/autotest_common.sh@10 -- # set +x 00:37:19.651 ************************************ 00:37:19.651 START TEST keyring_file 00:37:19.651 ************************************ 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:37:19.651 * Looking for test storage... 00:37:19.651 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1691 -- # lcov --version 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@333 -- # local ver1 ver1_l 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@334 -- # local ver2 ver2_l 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@336 -- # IFS=.-: 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@336 -- # read -ra ver1 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@337 -- # IFS=.-: 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@337 -- # read -ra ver2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@338 -- # local 'op=<' 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@340 -- # ver1_l=2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@341 -- # ver2_l=1 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@344 -- # case "$op" in 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@345 -- # : 1 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@364 -- # (( v = 0 )) 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@365 -- # decimal 1 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@353 -- # local d=1 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@355 -- # echo 1 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@365 -- # ver1[v]=1 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@366 -- # decimal 2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@353 -- # local d=2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@355 -- # echo 2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@366 -- # ver2[v]=2 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@368 -- # return 0 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:37:19.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:19.651 --rc genhtml_branch_coverage=1 00:37:19.651 --rc genhtml_function_coverage=1 00:37:19.651 --rc genhtml_legend=1 00:37:19.651 --rc geninfo_all_blocks=1 00:37:19.651 --rc geninfo_unexecuted_blocks=1 00:37:19.651 00:37:19.651 ' 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:37:19.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:19.651 --rc genhtml_branch_coverage=1 00:37:19.651 --rc genhtml_function_coverage=1 00:37:19.651 --rc genhtml_legend=1 00:37:19.651 --rc geninfo_all_blocks=1 00:37:19.651 --rc geninfo_unexecuted_blocks=1 00:37:19.651 00:37:19.651 ' 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:37:19.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:19.651 --rc genhtml_branch_coverage=1 00:37:19.651 --rc genhtml_function_coverage=1 00:37:19.651 --rc genhtml_legend=1 00:37:19.651 --rc geninfo_all_blocks=1 00:37:19.651 --rc geninfo_unexecuted_blocks=1 00:37:19.651 00:37:19.651 ' 00:37:19.651 17:58:27 keyring_file -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:37:19.651 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:19.651 --rc genhtml_branch_coverage=1 00:37:19.651 --rc genhtml_function_coverage=1 00:37:19.651 --rc genhtml_legend=1 00:37:19.651 --rc geninfo_all_blocks=1 00:37:19.651 --rc geninfo_unexecuted_blocks=1 00:37:19.651 00:37:19.651 ' 00:37:19.651 17:58:27 keyring_file -- keyring/file.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:37:19.651 17:58:27 keyring_file -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@7 -- # uname -s 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:37:19.651 17:58:27 keyring_file -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@15 -- # shopt -s extglob 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:37:19.651 17:58:27 keyring_file -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:37:19.652 17:58:27 keyring_file -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:37:19.652 17:58:27 keyring_file -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:19.652 17:58:27 keyring_file -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:19.652 17:58:27 keyring_file -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:19.652 17:58:27 keyring_file -- paths/export.sh@5 -- # export PATH 00:37:19.652 17:58:27 keyring_file -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@51 -- # : 0 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:37:19.652 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@55 -- # have_pci_nics=0 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@13 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@14 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@15 -- # key0=00112233445566778899aabbccddeeff 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@16 -- # key1=112233445566778899aabbccddeeff00 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@24 -- # trap cleanup EXIT 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@26 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@17 -- # name=key0 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@17 -- # digest=0 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@18 -- # mktemp 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.BMVOs3gndz 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@728 -- # local prefix key digest 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@730 -- # digest=0 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@731 -- # python - 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.BMVOs3gndz 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.BMVOs3gndz 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@26 -- # key0path=/tmp/tmp.BMVOs3gndz 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@27 -- # prep_key key1 112233445566778899aabbccddeeff00 0 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@17 -- # name=key1 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@17 -- # digest=0 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@18 -- # mktemp 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.7lTPjIV3S8 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@728 -- # local prefix key digest 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@730 -- # key=112233445566778899aabbccddeeff00 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@730 -- # digest=0 00:37:19.652 17:58:27 keyring_file -- nvmf/common.sh@731 -- # python - 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.7lTPjIV3S8 00:37:19.652 17:58:27 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.7lTPjIV3S8 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@27 -- # key1path=/tmp/tmp.7lTPjIV3S8 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@30 -- # tgtpid=377681 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@32 -- # waitforlisten 377681 00:37:19.652 17:58:27 keyring_file -- keyring/file.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:37:19.652 17:58:27 keyring_file -- common/autotest_common.sh@831 -- # '[' -z 377681 ']' 00:37:19.652 17:58:27 keyring_file -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:37:19.652 17:58:27 keyring_file -- common/autotest_common.sh@836 -- # local max_retries=100 00:37:19.652 17:58:27 keyring_file -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:37:19.652 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:37:19.652 17:58:27 keyring_file -- common/autotest_common.sh@840 -- # xtrace_disable 00:37:19.652 17:58:27 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:37:19.913 [2024-10-17 17:58:27.571362] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:37:19.913 [2024-10-17 17:58:27.571435] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid377681 ] 00:37:19.913 [2024-10-17 17:58:27.656397] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:19.913 [2024-10-17 17:58:27.710367] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:20.484 17:58:28 keyring_file -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:37:20.484 17:58:28 keyring_file -- common/autotest_common.sh@864 -- # return 0 00:37:20.484 17:58:28 keyring_file -- keyring/file.sh@33 -- # rpc_cmd 00:37:20.484 17:58:28 keyring_file -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:20.484 17:58:28 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:37:20.484 [2024-10-17 17:58:28.389484] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:37:20.746 null0 00:37:20.746 [2024-10-17 17:58:28.421535] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:37:20.746 [2024-10-17 17:58:28.421989] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:20.746 17:58:28 keyring_file -- keyring/file.sh@44 -- # NOT rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@653 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:37:20.746 [2024-10-17 17:58:28.453603] nvmf_rpc.c: 762:nvmf_rpc_listen_paused: *ERROR*: Listener already exists 00:37:20.746 request: 00:37:20.746 { 00:37:20.746 "nqn": "nqn.2016-06.io.spdk:cnode0", 00:37:20.746 "secure_channel": false, 00:37:20.746 "listen_address": { 00:37:20.746 "trtype": "tcp", 00:37:20.746 "traddr": "127.0.0.1", 00:37:20.746 "trsvcid": "4420" 00:37:20.746 }, 00:37:20.746 "method": "nvmf_subsystem_add_listener", 00:37:20.746 "req_id": 1 00:37:20.746 } 00:37:20.746 Got JSON-RPC error response 00:37:20.746 response: 00:37:20.746 { 00:37:20.746 "code": -32602, 00:37:20.746 "message": "Invalid parameters" 00:37:20.746 } 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:37:20.746 17:58:28 keyring_file -- keyring/file.sh@47 -- # bperfpid=377699 00:37:20.746 17:58:28 keyring_file -- keyring/file.sh@49 -- # waitforlisten 377699 /var/tmp/bperf.sock 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@831 -- # '[' -z 377699 ']' 00:37:20.746 17:58:28 keyring_file -- keyring/file.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@836 -- # local max_retries=100 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:20.746 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@840 -- # xtrace_disable 00:37:20.746 17:58:28 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:37:20.746 [2024-10-17 17:58:28.523125] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:37:20.746 [2024-10-17 17:58:28.523189] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid377699 ] 00:37:20.746 [2024-10-17 17:58:28.607125] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:20.746 [2024-10-17 17:58:28.660592] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:21.691 17:58:29 keyring_file -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:37:21.691 17:58:29 keyring_file -- common/autotest_common.sh@864 -- # return 0 00:37:21.691 17:58:29 keyring_file -- keyring/file.sh@50 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:21.691 17:58:29 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:21.691 17:58:29 keyring_file -- keyring/file.sh@51 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.7lTPjIV3S8 00:37:21.691 17:58:29 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.7lTPjIV3S8 00:37:21.952 17:58:29 keyring_file -- keyring/file.sh@52 -- # get_key key0 00:37:21.952 17:58:29 keyring_file -- keyring/file.sh@52 -- # jq -r .path 00:37:21.952 17:58:29 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:21.952 17:58:29 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:21.952 17:58:29 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:22.213 17:58:29 keyring_file -- keyring/file.sh@52 -- # [[ /tmp/tmp.BMVOs3gndz == \/\t\m\p\/\t\m\p\.\B\M\V\O\s\3\g\n\d\z ]] 00:37:22.213 17:58:29 keyring_file -- keyring/file.sh@53 -- # get_key key1 00:37:22.213 17:58:29 keyring_file -- keyring/file.sh@53 -- # jq -r .path 00:37:22.213 17:58:29 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:22.213 17:58:29 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:22.213 17:58:29 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:37:22.213 17:58:30 keyring_file -- keyring/file.sh@53 -- # [[ /tmp/tmp.7lTPjIV3S8 == \/\t\m\p\/\t\m\p\.\7\l\T\P\j\I\V\3\S\8 ]] 00:37:22.213 17:58:30 keyring_file -- keyring/file.sh@54 -- # get_refcnt key0 00:37:22.213 17:58:30 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:22.213 17:58:30 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:22.213 17:58:30 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:22.213 17:58:30 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:22.213 17:58:30 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:22.474 17:58:30 keyring_file -- keyring/file.sh@54 -- # (( 1 == 1 )) 00:37:22.474 17:58:30 keyring_file -- keyring/file.sh@55 -- # get_refcnt key1 00:37:22.474 17:58:30 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:37:22.474 17:58:30 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:22.474 17:58:30 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:22.474 17:58:30 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:22.474 17:58:30 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:37:22.736 17:58:30 keyring_file -- keyring/file.sh@55 -- # (( 1 == 1 )) 00:37:22.736 17:58:30 keyring_file -- keyring/file.sh@58 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:22.736 17:58:30 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:22.997 [2024-10-17 17:58:30.662750] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:37:22.997 nvme0n1 00:37:22.997 17:58:30 keyring_file -- keyring/file.sh@60 -- # get_refcnt key0 00:37:22.997 17:58:30 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:22.997 17:58:30 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:22.997 17:58:30 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:22.997 17:58:30 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:22.997 17:58:30 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:23.289 17:58:30 keyring_file -- keyring/file.sh@60 -- # (( 2 == 2 )) 00:37:23.289 17:58:30 keyring_file -- keyring/file.sh@61 -- # get_refcnt key1 00:37:23.289 17:58:30 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:37:23.289 17:58:30 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:23.289 17:58:30 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:23.289 17:58:30 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:23.289 17:58:30 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:37:23.289 17:58:31 keyring_file -- keyring/file.sh@61 -- # (( 1 == 1 )) 00:37:23.289 17:58:31 keyring_file -- keyring/file.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:37:23.560 Running I/O for 1 seconds... 00:37:24.501 19475.00 IOPS, 76.07 MiB/s 00:37:24.501 Latency(us) 00:37:24.501 [2024-10-17T15:58:32.420Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:24.501 Job: nvme0n1 (Core Mask 0x2, workload: randrw, percentage: 50, depth: 128, IO size: 4096) 00:37:24.501 nvme0n1 : 1.00 19528.81 76.28 0.00 0.00 6543.02 2457.60 14745.60 00:37:24.501 [2024-10-17T15:58:32.420Z] =================================================================================================================== 00:37:24.501 [2024-10-17T15:58:32.420Z] Total : 19528.81 76.28 0.00 0.00 6543.02 2457.60 14745.60 00:37:24.501 { 00:37:24.501 "results": [ 00:37:24.501 { 00:37:24.501 "job": "nvme0n1", 00:37:24.501 "core_mask": "0x2", 00:37:24.501 "workload": "randrw", 00:37:24.501 "percentage": 50, 00:37:24.501 "status": "finished", 00:37:24.501 "queue_depth": 128, 00:37:24.501 "io_size": 4096, 00:37:24.501 "runtime": 1.00385, 00:37:24.501 "iops": 19528.81406584649, 00:37:24.501 "mibps": 76.28442994471285, 00:37:24.501 "io_failed": 0, 00:37:24.501 "io_timeout": 0, 00:37:24.501 "avg_latency_us": 6543.01908454057, 00:37:24.501 "min_latency_us": 2457.6, 00:37:24.501 "max_latency_us": 14745.6 00:37:24.501 } 00:37:24.501 ], 00:37:24.501 "core_count": 1 00:37:24.501 } 00:37:24.501 17:58:32 keyring_file -- keyring/file.sh@65 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:37:24.501 17:58:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:37:24.763 17:58:32 keyring_file -- keyring/file.sh@66 -- # get_refcnt key0 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:24.763 17:58:32 keyring_file -- keyring/file.sh@66 -- # (( 1 == 1 )) 00:37:24.763 17:58:32 keyring_file -- keyring/file.sh@67 -- # get_refcnt key1 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:37:24.763 17:58:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:25.023 17:58:32 keyring_file -- keyring/file.sh@67 -- # (( 1 == 1 )) 00:37:25.023 17:58:32 keyring_file -- keyring/file.sh@70 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:37:25.023 17:58:32 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:37:25.023 17:58:32 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:37:25.023 17:58:32 keyring_file -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:37:25.023 17:58:32 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:25.023 17:58:32 keyring_file -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:37:25.023 17:58:32 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:25.023 17:58:32 keyring_file -- common/autotest_common.sh@653 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:37:25.023 17:58:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:37:25.282 [2024-10-17 17:58:32.945492] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:37:25.282 [2024-10-17 17:58:32.946260] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1344940 (107): Transport endpoint is not connected 00:37:25.282 [2024-10-17 17:58:32.947256] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1344940 (9): Bad file descriptor 00:37:25.282 [2024-10-17 17:58:32.948257] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:37:25.283 [2024-10-17 17:58:32.948265] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:37:25.283 [2024-10-17 17:58:32.948270] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:37:25.283 [2024-10-17 17:58:32.948276] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:37:25.283 request: 00:37:25.283 { 00:37:25.283 "name": "nvme0", 00:37:25.283 "trtype": "tcp", 00:37:25.283 "traddr": "127.0.0.1", 00:37:25.283 "adrfam": "ipv4", 00:37:25.283 "trsvcid": "4420", 00:37:25.283 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:25.283 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:25.283 "prchk_reftag": false, 00:37:25.283 "prchk_guard": false, 00:37:25.283 "hdgst": false, 00:37:25.283 "ddgst": false, 00:37:25.283 "psk": "key1", 00:37:25.283 "allow_unrecognized_csi": false, 00:37:25.283 "method": "bdev_nvme_attach_controller", 00:37:25.283 "req_id": 1 00:37:25.283 } 00:37:25.283 Got JSON-RPC error response 00:37:25.283 response: 00:37:25.283 { 00:37:25.283 "code": -5, 00:37:25.283 "message": "Input/output error" 00:37:25.283 } 00:37:25.283 17:58:32 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:37:25.283 17:58:32 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:37:25.283 17:58:32 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:37:25.283 17:58:32 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:37:25.283 17:58:32 keyring_file -- keyring/file.sh@72 -- # get_refcnt key0 00:37:25.283 17:58:32 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:25.283 17:58:32 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:25.283 17:58:32 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:25.283 17:58:32 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:25.283 17:58:32 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:25.283 17:58:33 keyring_file -- keyring/file.sh@72 -- # (( 1 == 1 )) 00:37:25.283 17:58:33 keyring_file -- keyring/file.sh@73 -- # get_refcnt key1 00:37:25.283 17:58:33 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:37:25.283 17:58:33 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:25.283 17:58:33 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:25.283 17:58:33 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:37:25.283 17:58:33 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:25.543 17:58:33 keyring_file -- keyring/file.sh@73 -- # (( 1 == 1 )) 00:37:25.543 17:58:33 keyring_file -- keyring/file.sh@76 -- # bperf_cmd keyring_file_remove_key key0 00:37:25.543 17:58:33 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:37:25.802 17:58:33 keyring_file -- keyring/file.sh@77 -- # bperf_cmd keyring_file_remove_key key1 00:37:25.802 17:58:33 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key1 00:37:25.802 17:58:33 keyring_file -- keyring/file.sh@78 -- # bperf_cmd keyring_get_keys 00:37:25.802 17:58:33 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:25.802 17:58:33 keyring_file -- keyring/file.sh@78 -- # jq length 00:37:26.062 17:58:33 keyring_file -- keyring/file.sh@78 -- # (( 0 == 0 )) 00:37:26.062 17:58:33 keyring_file -- keyring/file.sh@81 -- # chmod 0660 /tmp/tmp.BMVOs3gndz 00:37:26.063 17:58:33 keyring_file -- keyring/file.sh@82 -- # NOT bperf_cmd keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:26.063 17:58:33 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:37:26.063 17:58:33 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:26.063 17:58:33 keyring_file -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:37:26.063 17:58:33 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:26.063 17:58:33 keyring_file -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:37:26.063 17:58:33 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:26.063 17:58:33 keyring_file -- common/autotest_common.sh@653 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:26.063 17:58:33 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:26.063 [2024-10-17 17:58:33.976933] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.BMVOs3gndz': 0100660 00:37:26.063 [2024-10-17 17:58:33.976949] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:37:26.063 request: 00:37:26.063 { 00:37:26.063 "name": "key0", 00:37:26.063 "path": "/tmp/tmp.BMVOs3gndz", 00:37:26.063 "method": "keyring_file_add_key", 00:37:26.063 "req_id": 1 00:37:26.063 } 00:37:26.063 Got JSON-RPC error response 00:37:26.063 response: 00:37:26.063 { 00:37:26.063 "code": -1, 00:37:26.063 "message": "Operation not permitted" 00:37:26.063 } 00:37:26.323 17:58:33 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:37:26.323 17:58:33 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:37:26.323 17:58:33 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:37:26.323 17:58:33 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:37:26.323 17:58:33 keyring_file -- keyring/file.sh@85 -- # chmod 0600 /tmp/tmp.BMVOs3gndz 00:37:26.323 17:58:33 keyring_file -- keyring/file.sh@86 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:26.323 17:58:33 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.BMVOs3gndz 00:37:26.323 17:58:34 keyring_file -- keyring/file.sh@87 -- # rm -f /tmp/tmp.BMVOs3gndz 00:37:26.323 17:58:34 keyring_file -- keyring/file.sh@89 -- # get_refcnt key0 00:37:26.323 17:58:34 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:26.323 17:58:34 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:26.323 17:58:34 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:26.323 17:58:34 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:26.323 17:58:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:26.584 17:58:34 keyring_file -- keyring/file.sh@89 -- # (( 1 == 1 )) 00:37:26.584 17:58:34 keyring_file -- keyring/file.sh@91 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:26.584 17:58:34 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:37:26.584 17:58:34 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:26.584 17:58:34 keyring_file -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:37:26.584 17:58:34 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:26.584 17:58:34 keyring_file -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:37:26.584 17:58:34 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:26.584 17:58:34 keyring_file -- common/autotest_common.sh@653 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:26.584 17:58:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:26.584 [2024-10-17 17:58:34.502259] keyring.c: 31:keyring_file_check_path: *ERROR*: Could not stat key file '/tmp/tmp.BMVOs3gndz': No such file or directory 00:37:26.584 [2024-10-17 17:58:34.502273] nvme_tcp.c:2609:nvme_tcp_generate_tls_credentials: *ERROR*: Failed to obtain key 'key0': No such file or directory 00:37:26.584 [2024-10-17 17:58:34.502286] nvme.c: 682:nvme_ctrlr_probe: *ERROR*: Failed to construct NVMe controller for SSD: 127.0.0.1 00:37:26.584 [2024-10-17 17:58:34.502292] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, No such device 00:37:26.584 [2024-10-17 17:58:34.502301] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:37:26.584 [2024-10-17 17:58:34.502306] bdev_nvme.c:6436:spdk_bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 127.0.0.1) 00:37:26.845 request: 00:37:26.845 { 00:37:26.845 "name": "nvme0", 00:37:26.845 "trtype": "tcp", 00:37:26.845 "traddr": "127.0.0.1", 00:37:26.845 "adrfam": "ipv4", 00:37:26.845 "trsvcid": "4420", 00:37:26.845 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:26.845 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:26.845 "prchk_reftag": false, 00:37:26.845 "prchk_guard": false, 00:37:26.845 "hdgst": false, 00:37:26.845 "ddgst": false, 00:37:26.845 "psk": "key0", 00:37:26.845 "allow_unrecognized_csi": false, 00:37:26.845 "method": "bdev_nvme_attach_controller", 00:37:26.845 "req_id": 1 00:37:26.845 } 00:37:26.845 Got JSON-RPC error response 00:37:26.845 response: 00:37:26.845 { 00:37:26.845 "code": -19, 00:37:26.845 "message": "No such device" 00:37:26.845 } 00:37:26.845 17:58:34 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:37:26.845 17:58:34 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:37:26.845 17:58:34 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:37:26.845 17:58:34 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:37:26.845 17:58:34 keyring_file -- keyring/file.sh@93 -- # bperf_cmd keyring_file_remove_key key0 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:37:26.845 17:58:34 keyring_file -- keyring/file.sh@96 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@17 -- # name=key0 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@17 -- # digest=0 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@18 -- # mktemp 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.YV3adDWplk 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:37:26.845 17:58:34 keyring_file -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:37:26.845 17:58:34 keyring_file -- nvmf/common.sh@728 -- # local prefix key digest 00:37:26.845 17:58:34 keyring_file -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:37:26.845 17:58:34 keyring_file -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:37:26.845 17:58:34 keyring_file -- nvmf/common.sh@730 -- # digest=0 00:37:26.845 17:58:34 keyring_file -- nvmf/common.sh@731 -- # python - 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.YV3adDWplk 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.YV3adDWplk 00:37:26.845 17:58:34 keyring_file -- keyring/file.sh@96 -- # key0path=/tmp/tmp.YV3adDWplk 00:37:26.845 17:58:34 keyring_file -- keyring/file.sh@97 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.YV3adDWplk 00:37:26.845 17:58:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.YV3adDWplk 00:37:27.105 17:58:34 keyring_file -- keyring/file.sh@98 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:27.105 17:58:34 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:27.365 nvme0n1 00:37:27.365 17:58:35 keyring_file -- keyring/file.sh@100 -- # get_refcnt key0 00:37:27.365 17:58:35 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:27.365 17:58:35 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:27.365 17:58:35 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:27.365 17:58:35 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:27.365 17:58:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:27.625 17:58:35 keyring_file -- keyring/file.sh@100 -- # (( 2 == 2 )) 00:37:27.625 17:58:35 keyring_file -- keyring/file.sh@101 -- # bperf_cmd keyring_file_remove_key key0 00:37:27.625 17:58:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:37:27.625 17:58:35 keyring_file -- keyring/file.sh@102 -- # get_key key0 00:37:27.625 17:58:35 keyring_file -- keyring/file.sh@102 -- # jq -r .removed 00:37:27.625 17:58:35 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:27.625 17:58:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:27.625 17:58:35 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:27.885 17:58:35 keyring_file -- keyring/file.sh@102 -- # [[ true == \t\r\u\e ]] 00:37:27.885 17:58:35 keyring_file -- keyring/file.sh@103 -- # get_refcnt key0 00:37:27.885 17:58:35 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:27.885 17:58:35 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:27.885 17:58:35 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:27.885 17:58:35 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:27.885 17:58:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:28.145 17:58:35 keyring_file -- keyring/file.sh@103 -- # (( 1 == 1 )) 00:37:28.145 17:58:35 keyring_file -- keyring/file.sh@104 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:37:28.145 17:58:35 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:37:28.405 17:58:36 keyring_file -- keyring/file.sh@105 -- # bperf_cmd keyring_get_keys 00:37:28.405 17:58:36 keyring_file -- keyring/file.sh@105 -- # jq length 00:37:28.405 17:58:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:28.405 17:58:36 keyring_file -- keyring/file.sh@105 -- # (( 0 == 0 )) 00:37:28.405 17:58:36 keyring_file -- keyring/file.sh@108 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.YV3adDWplk 00:37:28.405 17:58:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.YV3adDWplk 00:37:28.666 17:58:36 keyring_file -- keyring/file.sh@109 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.7lTPjIV3S8 00:37:28.666 17:58:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.7lTPjIV3S8 00:37:28.666 17:58:36 keyring_file -- keyring/file.sh@110 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:28.666 17:58:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:37:28.926 nvme0n1 00:37:28.926 17:58:36 keyring_file -- keyring/file.sh@113 -- # bperf_cmd save_config 00:37:28.926 17:58:36 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock save_config 00:37:29.187 17:58:37 keyring_file -- keyring/file.sh@113 -- # config='{ 00:37:29.187 "subsystems": [ 00:37:29.187 { 00:37:29.187 "subsystem": "keyring", 00:37:29.187 "config": [ 00:37:29.187 { 00:37:29.187 "method": "keyring_file_add_key", 00:37:29.187 "params": { 00:37:29.187 "name": "key0", 00:37:29.187 "path": "/tmp/tmp.YV3adDWplk" 00:37:29.187 } 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "method": "keyring_file_add_key", 00:37:29.187 "params": { 00:37:29.187 "name": "key1", 00:37:29.187 "path": "/tmp/tmp.7lTPjIV3S8" 00:37:29.187 } 00:37:29.187 } 00:37:29.187 ] 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "subsystem": "iobuf", 00:37:29.187 "config": [ 00:37:29.187 { 00:37:29.187 "method": "iobuf_set_options", 00:37:29.187 "params": { 00:37:29.187 "small_pool_count": 8192, 00:37:29.187 "large_pool_count": 1024, 00:37:29.187 "small_bufsize": 8192, 00:37:29.187 "large_bufsize": 135168 00:37:29.187 } 00:37:29.187 } 00:37:29.187 ] 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "subsystem": "sock", 00:37:29.187 "config": [ 00:37:29.187 { 00:37:29.187 "method": "sock_set_default_impl", 00:37:29.187 "params": { 00:37:29.187 "impl_name": "posix" 00:37:29.187 } 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "method": "sock_impl_set_options", 00:37:29.187 "params": { 00:37:29.187 "impl_name": "ssl", 00:37:29.187 "recv_buf_size": 4096, 00:37:29.187 "send_buf_size": 4096, 00:37:29.187 "enable_recv_pipe": true, 00:37:29.187 "enable_quickack": false, 00:37:29.187 "enable_placement_id": 0, 00:37:29.187 "enable_zerocopy_send_server": true, 00:37:29.187 "enable_zerocopy_send_client": false, 00:37:29.187 "zerocopy_threshold": 0, 00:37:29.187 "tls_version": 0, 00:37:29.187 "enable_ktls": false 00:37:29.187 } 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "method": "sock_impl_set_options", 00:37:29.187 "params": { 00:37:29.187 "impl_name": "posix", 00:37:29.187 "recv_buf_size": 2097152, 00:37:29.187 "send_buf_size": 2097152, 00:37:29.187 "enable_recv_pipe": true, 00:37:29.187 "enable_quickack": false, 00:37:29.187 "enable_placement_id": 0, 00:37:29.187 "enable_zerocopy_send_server": true, 00:37:29.187 "enable_zerocopy_send_client": false, 00:37:29.187 "zerocopy_threshold": 0, 00:37:29.187 "tls_version": 0, 00:37:29.187 "enable_ktls": false 00:37:29.187 } 00:37:29.187 } 00:37:29.187 ] 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "subsystem": "vmd", 00:37:29.187 "config": [] 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "subsystem": "accel", 00:37:29.187 "config": [ 00:37:29.187 { 00:37:29.187 "method": "accel_set_options", 00:37:29.187 "params": { 00:37:29.187 "small_cache_size": 128, 00:37:29.187 "large_cache_size": 16, 00:37:29.187 "task_count": 2048, 00:37:29.187 "sequence_count": 2048, 00:37:29.187 "buf_count": 2048 00:37:29.187 } 00:37:29.187 } 00:37:29.187 ] 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "subsystem": "bdev", 00:37:29.187 "config": [ 00:37:29.187 { 00:37:29.187 "method": "bdev_set_options", 00:37:29.187 "params": { 00:37:29.187 "bdev_io_pool_size": 65535, 00:37:29.187 "bdev_io_cache_size": 256, 00:37:29.187 "bdev_auto_examine": true, 00:37:29.187 "iobuf_small_cache_size": 128, 00:37:29.187 "iobuf_large_cache_size": 16 00:37:29.187 } 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "method": "bdev_raid_set_options", 00:37:29.187 "params": { 00:37:29.187 "process_window_size_kb": 1024, 00:37:29.187 "process_max_bandwidth_mb_sec": 0 00:37:29.187 } 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "method": "bdev_iscsi_set_options", 00:37:29.187 "params": { 00:37:29.187 "timeout_sec": 30 00:37:29.187 } 00:37:29.187 }, 00:37:29.187 { 00:37:29.187 "method": "bdev_nvme_set_options", 00:37:29.187 "params": { 00:37:29.187 "action_on_timeout": "none", 00:37:29.187 "timeout_us": 0, 00:37:29.187 "timeout_admin_us": 0, 00:37:29.187 "keep_alive_timeout_ms": 10000, 00:37:29.187 "arbitration_burst": 0, 00:37:29.187 "low_priority_weight": 0, 00:37:29.187 "medium_priority_weight": 0, 00:37:29.187 "high_priority_weight": 0, 00:37:29.187 "nvme_adminq_poll_period_us": 10000, 00:37:29.187 "nvme_ioq_poll_period_us": 0, 00:37:29.187 "io_queue_requests": 512, 00:37:29.187 "delay_cmd_submit": true, 00:37:29.187 "transport_retry_count": 4, 00:37:29.187 "bdev_retry_count": 3, 00:37:29.188 "transport_ack_timeout": 0, 00:37:29.188 "ctrlr_loss_timeout_sec": 0, 00:37:29.188 "reconnect_delay_sec": 0, 00:37:29.188 "fast_io_fail_timeout_sec": 0, 00:37:29.188 "disable_auto_failback": false, 00:37:29.188 "generate_uuids": false, 00:37:29.188 "transport_tos": 0, 00:37:29.188 "nvme_error_stat": false, 00:37:29.188 "rdma_srq_size": 0, 00:37:29.188 "io_path_stat": false, 00:37:29.188 "allow_accel_sequence": false, 00:37:29.188 "rdma_max_cq_size": 0, 00:37:29.188 "rdma_cm_event_timeout_ms": 0, 00:37:29.188 "dhchap_digests": [ 00:37:29.188 "sha256", 00:37:29.188 "sha384", 00:37:29.188 "sha512" 00:37:29.188 ], 00:37:29.188 "dhchap_dhgroups": [ 00:37:29.188 "null", 00:37:29.188 "ffdhe2048", 00:37:29.188 "ffdhe3072", 00:37:29.188 "ffdhe4096", 00:37:29.188 "ffdhe6144", 00:37:29.188 "ffdhe8192" 00:37:29.188 ], 00:37:29.188 "rdma_umr_per_io": false 00:37:29.188 } 00:37:29.188 }, 00:37:29.188 { 00:37:29.188 "method": "bdev_nvme_attach_controller", 00:37:29.188 "params": { 00:37:29.188 "name": "nvme0", 00:37:29.188 "trtype": "TCP", 00:37:29.188 "adrfam": "IPv4", 00:37:29.188 "traddr": "127.0.0.1", 00:37:29.188 "trsvcid": "4420", 00:37:29.188 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:29.188 "prchk_reftag": false, 00:37:29.188 "prchk_guard": false, 00:37:29.188 "ctrlr_loss_timeout_sec": 0, 00:37:29.188 "reconnect_delay_sec": 0, 00:37:29.188 "fast_io_fail_timeout_sec": 0, 00:37:29.188 "psk": "key0", 00:37:29.188 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:29.188 "hdgst": false, 00:37:29.188 "ddgst": false, 00:37:29.188 "multipath": "multipath" 00:37:29.188 } 00:37:29.188 }, 00:37:29.188 { 00:37:29.188 "method": "bdev_nvme_set_hotplug", 00:37:29.188 "params": { 00:37:29.188 "period_us": 100000, 00:37:29.188 "enable": false 00:37:29.188 } 00:37:29.188 }, 00:37:29.188 { 00:37:29.188 "method": "bdev_wait_for_examine" 00:37:29.188 } 00:37:29.188 ] 00:37:29.188 }, 00:37:29.188 { 00:37:29.188 "subsystem": "nbd", 00:37:29.188 "config": [] 00:37:29.188 } 00:37:29.188 ] 00:37:29.188 }' 00:37:29.188 17:58:37 keyring_file -- keyring/file.sh@115 -- # killprocess 377699 00:37:29.188 17:58:37 keyring_file -- common/autotest_common.sh@950 -- # '[' -z 377699 ']' 00:37:29.188 17:58:37 keyring_file -- common/autotest_common.sh@954 -- # kill -0 377699 00:37:29.188 17:58:37 keyring_file -- common/autotest_common.sh@955 -- # uname 00:37:29.188 17:58:37 keyring_file -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:29.188 17:58:37 keyring_file -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 377699 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@968 -- # echo 'killing process with pid 377699' 00:37:29.449 killing process with pid 377699 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@969 -- # kill 377699 00:37:29.449 Received shutdown signal, test time was about 1.000000 seconds 00:37:29.449 00:37:29.449 Latency(us) 00:37:29.449 [2024-10-17T15:58:37.368Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:29.449 [2024-10-17T15:58:37.368Z] =================================================================================================================== 00:37:29.449 [2024-10-17T15:58:37.368Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@974 -- # wait 377699 00:37:29.449 17:58:37 keyring_file -- keyring/file.sh@118 -- # bperfpid=379512 00:37:29.449 17:58:37 keyring_file -- keyring/file.sh@120 -- # waitforlisten 379512 /var/tmp/bperf.sock 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@831 -- # '[' -z 379512 ']' 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:29.449 17:58:37 keyring_file -- keyring/file.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z -c /dev/fd/63 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@836 -- # local max_retries=100 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:29.449 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:29.449 17:58:37 keyring_file -- common/autotest_common.sh@840 -- # xtrace_disable 00:37:29.449 17:58:37 keyring_file -- keyring/file.sh@116 -- # echo '{ 00:37:29.449 "subsystems": [ 00:37:29.449 { 00:37:29.449 "subsystem": "keyring", 00:37:29.449 "config": [ 00:37:29.449 { 00:37:29.449 "method": "keyring_file_add_key", 00:37:29.449 "params": { 00:37:29.449 "name": "key0", 00:37:29.449 "path": "/tmp/tmp.YV3adDWplk" 00:37:29.449 } 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "method": "keyring_file_add_key", 00:37:29.449 "params": { 00:37:29.449 "name": "key1", 00:37:29.449 "path": "/tmp/tmp.7lTPjIV3S8" 00:37:29.449 } 00:37:29.449 } 00:37:29.449 ] 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "subsystem": "iobuf", 00:37:29.449 "config": [ 00:37:29.449 { 00:37:29.449 "method": "iobuf_set_options", 00:37:29.449 "params": { 00:37:29.449 "small_pool_count": 8192, 00:37:29.449 "large_pool_count": 1024, 00:37:29.449 "small_bufsize": 8192, 00:37:29.449 "large_bufsize": 135168 00:37:29.449 } 00:37:29.449 } 00:37:29.449 ] 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "subsystem": "sock", 00:37:29.449 "config": [ 00:37:29.449 { 00:37:29.449 "method": "sock_set_default_impl", 00:37:29.449 "params": { 00:37:29.449 "impl_name": "posix" 00:37:29.449 } 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "method": "sock_impl_set_options", 00:37:29.449 "params": { 00:37:29.449 "impl_name": "ssl", 00:37:29.449 "recv_buf_size": 4096, 00:37:29.449 "send_buf_size": 4096, 00:37:29.449 "enable_recv_pipe": true, 00:37:29.449 "enable_quickack": false, 00:37:29.449 "enable_placement_id": 0, 00:37:29.449 "enable_zerocopy_send_server": true, 00:37:29.449 "enable_zerocopy_send_client": false, 00:37:29.449 "zerocopy_threshold": 0, 00:37:29.449 "tls_version": 0, 00:37:29.449 "enable_ktls": false 00:37:29.449 } 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "method": "sock_impl_set_options", 00:37:29.449 "params": { 00:37:29.449 "impl_name": "posix", 00:37:29.449 "recv_buf_size": 2097152, 00:37:29.449 "send_buf_size": 2097152, 00:37:29.449 "enable_recv_pipe": true, 00:37:29.449 "enable_quickack": false, 00:37:29.449 "enable_placement_id": 0, 00:37:29.449 "enable_zerocopy_send_server": true, 00:37:29.449 "enable_zerocopy_send_client": false, 00:37:29.449 "zerocopy_threshold": 0, 00:37:29.449 "tls_version": 0, 00:37:29.449 "enable_ktls": false 00:37:29.449 } 00:37:29.449 } 00:37:29.449 ] 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "subsystem": "vmd", 00:37:29.449 "config": [] 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "subsystem": "accel", 00:37:29.449 "config": [ 00:37:29.449 { 00:37:29.449 "method": "accel_set_options", 00:37:29.449 "params": { 00:37:29.449 "small_cache_size": 128, 00:37:29.449 "large_cache_size": 16, 00:37:29.449 "task_count": 2048, 00:37:29.449 "sequence_count": 2048, 00:37:29.449 "buf_count": 2048 00:37:29.449 } 00:37:29.449 } 00:37:29.449 ] 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "subsystem": "bdev", 00:37:29.449 "config": [ 00:37:29.449 { 00:37:29.449 "method": "bdev_set_options", 00:37:29.449 "params": { 00:37:29.449 "bdev_io_pool_size": 65535, 00:37:29.449 "bdev_io_cache_size": 256, 00:37:29.449 "bdev_auto_examine": true, 00:37:29.449 "iobuf_small_cache_size": 128, 00:37:29.449 "iobuf_large_cache_size": 16 00:37:29.449 } 00:37:29.449 }, 00:37:29.449 { 00:37:29.449 "method": "bdev_raid_set_options", 00:37:29.449 "params": { 00:37:29.450 "process_window_size_kb": 1024, 00:37:29.450 "process_max_bandwidth_mb_sec": 0 00:37:29.450 } 00:37:29.450 }, 00:37:29.450 { 00:37:29.450 "method": "bdev_iscsi_set_options", 00:37:29.450 "params": { 00:37:29.450 "timeout_sec": 30 00:37:29.450 } 00:37:29.450 }, 00:37:29.450 { 00:37:29.450 "method": "bdev_nvme_set_options", 00:37:29.450 "params": { 00:37:29.450 "action_on_timeout": "none", 00:37:29.450 "timeout_us": 0, 00:37:29.450 "timeout_admin_us": 0, 00:37:29.450 "keep_alive_timeout_ms": 10000, 00:37:29.450 "arbitration_burst": 0, 00:37:29.450 "low_priority_weight": 0, 00:37:29.450 "medium_priority_weight": 0, 00:37:29.450 "high_priority_weight": 0, 00:37:29.450 "nvme_adminq_poll_period_us": 10000, 00:37:29.450 "nvme_ioq_poll_period_us": 0, 00:37:29.450 "io_queue_requests": 512, 00:37:29.450 "delay_cmd_submit": true, 00:37:29.450 "transport_retry_count": 4, 00:37:29.450 "bdev_retry_count": 3, 00:37:29.450 "transport_ack_timeout": 0, 00:37:29.450 "ctrlr_loss_timeout_sec": 0, 00:37:29.450 "reconnect_delay_sec": 0, 00:37:29.450 "fast_io_fail_timeout_sec": 0, 00:37:29.450 "disable_auto_failback": false, 00:37:29.450 "generate_uuids": false, 00:37:29.450 "transport_tos": 0, 00:37:29.450 "nvme_error_stat": false, 00:37:29.450 "rdma_srq_size": 0, 00:37:29.450 "io_path_stat": false, 00:37:29.450 17:58:37 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:37:29.450 "allow_accel_sequence": false, 00:37:29.450 "rdma_max_cq_size": 0, 00:37:29.450 "rdma_cm_event_timeout_ms": 0, 00:37:29.450 "dhchap_digests": [ 00:37:29.450 "sha256", 00:37:29.450 "sha384", 00:37:29.450 "sha512" 00:37:29.450 ], 00:37:29.450 "dhchap_dhgroups": [ 00:37:29.450 "null", 00:37:29.450 "ffdhe2048", 00:37:29.450 "ffdhe3072", 00:37:29.450 "ffdhe4096", 00:37:29.450 "ffdhe6144", 00:37:29.450 "ffdhe8192" 00:37:29.450 ], 00:37:29.450 "rdma_umr_per_io": false 00:37:29.450 } 00:37:29.450 }, 00:37:29.450 { 00:37:29.450 "method": "bdev_nvme_attach_controller", 00:37:29.450 "params": { 00:37:29.450 "name": "nvme0", 00:37:29.450 "trtype": "TCP", 00:37:29.450 "adrfam": "IPv4", 00:37:29.450 "traddr": "127.0.0.1", 00:37:29.450 "trsvcid": "4420", 00:37:29.450 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:29.450 "prchk_reftag": false, 00:37:29.450 "prchk_guard": false, 00:37:29.450 "ctrlr_loss_timeout_sec": 0, 00:37:29.450 "reconnect_delay_sec": 0, 00:37:29.450 "fast_io_fail_timeout_sec": 0, 00:37:29.450 "psk": "key0", 00:37:29.450 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:29.450 "hdgst": false, 00:37:29.450 "ddgst": false, 00:37:29.450 "multipath": "multipath" 00:37:29.450 } 00:37:29.450 }, 00:37:29.450 { 00:37:29.450 "method": "bdev_nvme_set_hotplug", 00:37:29.450 "params": { 00:37:29.450 "period_us": 100000, 00:37:29.450 "enable": false 00:37:29.450 } 00:37:29.450 }, 00:37:29.450 { 00:37:29.450 "method": "bdev_wait_for_examine" 00:37:29.450 } 00:37:29.450 ] 00:37:29.450 }, 00:37:29.450 { 00:37:29.450 "subsystem": "nbd", 00:37:29.450 "config": [] 00:37:29.450 } 00:37:29.450 ] 00:37:29.450 }' 00:37:29.450 [2024-10-17 17:58:37.271360] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:37:29.450 [2024-10-17 17:58:37.271418] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid379512 ] 00:37:29.450 [2024-10-17 17:58:37.347302] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:29.710 [2024-10-17 17:58:37.376437] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:29.710 [2024-10-17 17:58:37.519415] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:37:30.279 17:58:38 keyring_file -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:37:30.279 17:58:38 keyring_file -- common/autotest_common.sh@864 -- # return 0 00:37:30.279 17:58:38 keyring_file -- keyring/file.sh@121 -- # bperf_cmd keyring_get_keys 00:37:30.279 17:58:38 keyring_file -- keyring/file.sh@121 -- # jq length 00:37:30.279 17:58:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:30.540 17:58:38 keyring_file -- keyring/file.sh@121 -- # (( 2 == 2 )) 00:37:30.540 17:58:38 keyring_file -- keyring/file.sh@122 -- # get_refcnt key0 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:30.540 17:58:38 keyring_file -- keyring/file.sh@122 -- # (( 2 == 2 )) 00:37:30.540 17:58:38 keyring_file -- keyring/file.sh@123 -- # get_refcnt key1 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:37:30.540 17:58:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:30.800 17:58:38 keyring_file -- keyring/file.sh@123 -- # (( 1 == 1 )) 00:37:30.800 17:58:38 keyring_file -- keyring/file.sh@124 -- # bperf_cmd bdev_nvme_get_controllers 00:37:30.800 17:58:38 keyring_file -- keyring/file.sh@124 -- # jq -r '.[].name' 00:37:30.800 17:58:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_get_controllers 00:37:31.060 17:58:38 keyring_file -- keyring/file.sh@124 -- # [[ nvme0 == nvme0 ]] 00:37:31.060 17:58:38 keyring_file -- keyring/file.sh@1 -- # cleanup 00:37:31.060 17:58:38 keyring_file -- keyring/file.sh@19 -- # rm -f /tmp/tmp.YV3adDWplk /tmp/tmp.7lTPjIV3S8 00:37:31.060 17:58:38 keyring_file -- keyring/file.sh@20 -- # killprocess 379512 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@950 -- # '[' -z 379512 ']' 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@954 -- # kill -0 379512 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@955 -- # uname 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 379512 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@968 -- # echo 'killing process with pid 379512' 00:37:31.060 killing process with pid 379512 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@969 -- # kill 379512 00:37:31.060 Received shutdown signal, test time was about 1.000000 seconds 00:37:31.060 00:37:31.060 Latency(us) 00:37:31.060 [2024-10-17T15:58:38.979Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:31.060 [2024-10-17T15:58:38.979Z] =================================================================================================================== 00:37:31.060 [2024-10-17T15:58:38.979Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@974 -- # wait 379512 00:37:31.060 17:58:38 keyring_file -- keyring/file.sh@21 -- # killprocess 377681 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@950 -- # '[' -z 377681 ']' 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@954 -- # kill -0 377681 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@955 -- # uname 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:31.060 17:58:38 keyring_file -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 377681 00:37:31.320 17:58:38 keyring_file -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:37:31.320 17:58:38 keyring_file -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:37:31.320 17:58:38 keyring_file -- common/autotest_common.sh@968 -- # echo 'killing process with pid 377681' 00:37:31.320 killing process with pid 377681 00:37:31.320 17:58:38 keyring_file -- common/autotest_common.sh@969 -- # kill 377681 00:37:31.320 17:58:38 keyring_file -- common/autotest_common.sh@974 -- # wait 377681 00:37:31.320 00:37:31.320 real 0m12.040s 00:37:31.320 user 0m28.988s 00:37:31.320 sys 0m2.763s 00:37:31.320 17:58:39 keyring_file -- common/autotest_common.sh@1126 -- # xtrace_disable 00:37:31.320 17:58:39 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:37:31.320 ************************************ 00:37:31.320 END TEST keyring_file 00:37:31.320 ************************************ 00:37:31.320 17:58:39 -- spdk/autotest.sh@289 -- # [[ y == y ]] 00:37:31.320 17:58:39 -- spdk/autotest.sh@290 -- # run_test keyring_linux /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:37:31.320 17:58:39 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:37:31.320 17:58:39 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:37:31.320 17:58:39 -- common/autotest_common.sh@10 -- # set +x 00:37:31.580 ************************************ 00:37:31.580 START TEST keyring_linux 00:37:31.580 ************************************ 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:37:31.580 Joined session keyring: 266562950 00:37:31.580 * Looking for test storage... 00:37:31.580 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1691 -- # lcov --version 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@333 -- # local ver1 ver1_l 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@334 -- # local ver2 ver2_l 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@336 -- # IFS=.-: 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@336 -- # read -ra ver1 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@337 -- # IFS=.-: 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@337 -- # read -ra ver2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@338 -- # local 'op=<' 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@340 -- # ver1_l=2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@341 -- # ver2_l=1 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@344 -- # case "$op" in 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@345 -- # : 1 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@364 -- # (( v = 0 )) 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@365 -- # decimal 1 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@353 -- # local d=1 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@355 -- # echo 1 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@365 -- # ver1[v]=1 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@366 -- # decimal 2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@353 -- # local d=2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@355 -- # echo 2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@366 -- # ver2[v]=2 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@368 -- # return 0 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:37:31.580 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:31.580 --rc genhtml_branch_coverage=1 00:37:31.580 --rc genhtml_function_coverage=1 00:37:31.580 --rc genhtml_legend=1 00:37:31.580 --rc geninfo_all_blocks=1 00:37:31.580 --rc geninfo_unexecuted_blocks=1 00:37:31.580 00:37:31.580 ' 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:37:31.580 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:31.580 --rc genhtml_branch_coverage=1 00:37:31.580 --rc genhtml_function_coverage=1 00:37:31.580 --rc genhtml_legend=1 00:37:31.580 --rc geninfo_all_blocks=1 00:37:31.580 --rc geninfo_unexecuted_blocks=1 00:37:31.580 00:37:31.580 ' 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:37:31.580 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:31.580 --rc genhtml_branch_coverage=1 00:37:31.580 --rc genhtml_function_coverage=1 00:37:31.580 --rc genhtml_legend=1 00:37:31.580 --rc geninfo_all_blocks=1 00:37:31.580 --rc geninfo_unexecuted_blocks=1 00:37:31.580 00:37:31.580 ' 00:37:31.580 17:58:39 keyring_linux -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:37:31.580 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:31.580 --rc genhtml_branch_coverage=1 00:37:31.580 --rc genhtml_function_coverage=1 00:37:31.580 --rc genhtml_legend=1 00:37:31.580 --rc geninfo_all_blocks=1 00:37:31.580 --rc geninfo_unexecuted_blocks=1 00:37:31.580 00:37:31.580 ' 00:37:31.580 17:58:39 keyring_linux -- keyring/linux.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:37:31.580 17:58:39 keyring_linux -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@7 -- # uname -s 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@18 -- # NVME_HOSTID=008c5ac1-5feb-ec11-9bc7-a4bf019282a6 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@15 -- # shopt -s extglob 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:37:31.580 17:58:39 keyring_linux -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:37:31.580 17:58:39 keyring_linux -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:31.580 17:58:39 keyring_linux -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:31.580 17:58:39 keyring_linux -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:31.580 17:58:39 keyring_linux -- paths/export.sh@5 -- # export PATH 00:37:31.580 17:58:39 keyring_linux -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@51 -- # : 0 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:37:31.580 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:37:31.580 17:58:39 keyring_linux -- nvmf/common.sh@55 -- # have_pci_nics=0 00:37:31.580 17:58:39 keyring_linux -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@11 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@12 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@13 -- # key0=00112233445566778899aabbccddeeff 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@14 -- # key1=112233445566778899aabbccddeeff00 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@45 -- # trap cleanup EXIT 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@47 -- # prep_key key0 00112233445566778899aabbccddeeff 0 /tmp/:spdk-test:key0 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@17 -- # name=key0 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key0 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@728 -- # local prefix key digest 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@730 -- # digest=0 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@731 -- # python - 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key0 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key0 00:37:31.841 /tmp/:spdk-test:key0 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@48 -- # prep_key key1 112233445566778899aabbccddeeff00 0 /tmp/:spdk-test:key1 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@17 -- # name=key1 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key1 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@728 -- # local prefix key digest 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@730 -- # key=112233445566778899aabbccddeeff00 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@730 -- # digest=0 00:37:31.841 17:58:39 keyring_linux -- nvmf/common.sh@731 -- # python - 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key1 00:37:31.841 17:58:39 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key1 00:37:31.841 /tmp/:spdk-test:key1 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@51 -- # tgtpid=379963 00:37:31.841 17:58:39 keyring_linux -- keyring/linux.sh@53 -- # waitforlisten 379963 00:37:31.841 17:58:39 keyring_linux -- common/autotest_common.sh@831 -- # '[' -z 379963 ']' 00:37:31.841 17:58:39 keyring_linux -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:37:31.841 17:58:39 keyring_linux -- common/autotest_common.sh@836 -- # local max_retries=100 00:37:31.841 17:58:39 keyring_linux -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:37:31.841 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:37:31.841 17:58:39 keyring_linux -- common/autotest_common.sh@840 -- # xtrace_disable 00:37:31.841 17:58:39 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:37:31.841 [2024-10-17 17:58:39.634655] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:37:31.841 [2024-10-17 17:58:39.634722] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid379963 ] 00:37:31.841 [2024-10-17 17:58:39.716387] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:31.841 [2024-10-17 17:58:39.751752] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:32.101 17:58:39 keyring_linux -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:37:32.101 17:58:39 keyring_linux -- common/autotest_common.sh@864 -- # return 0 00:37:32.101 17:58:39 keyring_linux -- keyring/linux.sh@54 -- # rpc_cmd 00:37:32.101 17:58:39 keyring_linux -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:32.101 17:58:39 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:37:32.101 [2024-10-17 17:58:39.942553] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:37:32.101 null0 00:37:32.101 [2024-10-17 17:58:39.974613] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:37:32.101 [2024-10-17 17:58:39.974976] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:37:32.101 17:58:39 keyring_linux -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:32.101 17:58:39 keyring_linux -- keyring/linux.sh@66 -- # keyctl add user :spdk-test:key0 NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: @s 00:37:32.101 435014426 00:37:32.101 17:58:39 keyring_linux -- keyring/linux.sh@67 -- # keyctl add user :spdk-test:key1 NVMeTLSkey-1:00:MTEyMjMzNDQ1NTY2Nzc4ODk5YWFiYmNjZGRlZWZmMDA6CPcs: @s 00:37:32.101 943285474 00:37:32.101 17:58:40 keyring_linux -- keyring/linux.sh@70 -- # bperfpid=380121 00:37:32.101 17:58:40 keyring_linux -- keyring/linux.sh@72 -- # waitforlisten 380121 /var/tmp/bperf.sock 00:37:32.101 17:58:40 keyring_linux -- keyring/linux.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randread -t 1 -m 2 -r /var/tmp/bperf.sock -z --wait-for-rpc 00:37:32.101 17:58:40 keyring_linux -- common/autotest_common.sh@831 -- # '[' -z 380121 ']' 00:37:32.101 17:58:40 keyring_linux -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:32.101 17:58:40 keyring_linux -- common/autotest_common.sh@836 -- # local max_retries=100 00:37:32.101 17:58:40 keyring_linux -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:32.101 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:32.101 17:58:40 keyring_linux -- common/autotest_common.sh@840 -- # xtrace_disable 00:37:32.101 17:58:40 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:37:32.361 [2024-10-17 17:58:40.051434] Starting SPDK v25.01-pre git sha1 18d1cad7b / DPDK 24.03.0 initialization... 00:37:32.362 [2024-10-17 17:58:40.051489] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid380121 ] 00:37:32.362 [2024-10-17 17:58:40.127802] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:32.362 [2024-10-17 17:58:40.157932] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:32.362 17:58:40 keyring_linux -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:37:32.362 17:58:40 keyring_linux -- common/autotest_common.sh@864 -- # return 0 00:37:32.362 17:58:40 keyring_linux -- keyring/linux.sh@73 -- # bperf_cmd keyring_linux_set_options --enable 00:37:32.362 17:58:40 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_linux_set_options --enable 00:37:32.621 17:58:40 keyring_linux -- keyring/linux.sh@74 -- # bperf_cmd framework_start_init 00:37:32.621 17:58:40 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:37:32.882 17:58:40 keyring_linux -- keyring/linux.sh@75 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:37:32.882 17:58:40 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:37:32.882 [2024-10-17 17:58:40.740646] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:37:33.142 nvme0n1 00:37:33.142 17:58:40 keyring_linux -- keyring/linux.sh@77 -- # check_keys 1 :spdk-test:key0 00:37:33.142 17:58:40 keyring_linux -- keyring/linux.sh@19 -- # local count=1 name=:spdk-test:key0 00:37:33.142 17:58:40 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:37:33.142 17:58:40 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:37:33.142 17:58:40 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:37:33.142 17:58:40 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:33.142 17:58:41 keyring_linux -- keyring/linux.sh@22 -- # (( 1 == count )) 00:37:33.142 17:58:41 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:37:33.142 17:58:41 keyring_linux -- keyring/linux.sh@25 -- # get_key :spdk-test:key0 00:37:33.142 17:58:41 keyring_linux -- keyring/linux.sh@25 -- # jq -r .sn 00:37:33.142 17:58:41 keyring_linux -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:37:33.142 17:58:41 keyring_linux -- keyring/common.sh@10 -- # jq '.[] | select(.name == ":spdk-test:key0")' 00:37:33.142 17:58:41 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:33.403 17:58:41 keyring_linux -- keyring/linux.sh@25 -- # sn=435014426 00:37:33.403 17:58:41 keyring_linux -- keyring/linux.sh@26 -- # get_keysn :spdk-test:key0 00:37:33.403 17:58:41 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:37:33.403 17:58:41 keyring_linux -- keyring/linux.sh@26 -- # [[ 435014426 == \4\3\5\0\1\4\4\2\6 ]] 00:37:33.403 17:58:41 keyring_linux -- keyring/linux.sh@27 -- # keyctl print 435014426 00:37:33.403 17:58:41 keyring_linux -- keyring/linux.sh@27 -- # [[ NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: == \N\V\M\e\T\L\S\k\e\y\-\1\:\0\0\:\M\D\A\x\M\T\I\y\M\z\M\0\N\D\U\1\N\j\Y\3\N\z\g\4\O\T\l\h\Y\W\J\i\Y\2\N\k\Z\G\V\l\Z\m\Z\w\J\E\i\Q\: ]] 00:37:33.403 17:58:41 keyring_linux -- keyring/linux.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:37:33.403 Running I/O for 1 seconds... 00:37:34.789 24449.00 IOPS, 95.50 MiB/s 00:37:34.789 Latency(us) 00:37:34.789 [2024-10-17T15:58:42.708Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:34.789 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:37:34.789 nvme0n1 : 1.01 24449.74 95.51 0.00 0.00 5219.86 4314.45 12397.23 00:37:34.789 [2024-10-17T15:58:42.708Z] =================================================================================================================== 00:37:34.789 [2024-10-17T15:58:42.708Z] Total : 24449.74 95.51 0.00 0.00 5219.86 4314.45 12397.23 00:37:34.789 { 00:37:34.789 "results": [ 00:37:34.789 { 00:37:34.789 "job": "nvme0n1", 00:37:34.789 "core_mask": "0x2", 00:37:34.789 "workload": "randread", 00:37:34.789 "status": "finished", 00:37:34.789 "queue_depth": 128, 00:37:34.789 "io_size": 4096, 00:37:34.789 "runtime": 1.005205, 00:37:34.789 "iops": 24449.739107943155, 00:37:34.789 "mibps": 95.50679339040295, 00:37:34.789 "io_failed": 0, 00:37:34.789 "io_timeout": 0, 00:37:34.789 "avg_latency_us": 5219.864825378741, 00:37:34.789 "min_latency_us": 4314.453333333333, 00:37:34.789 "max_latency_us": 12397.226666666667 00:37:34.789 } 00:37:34.789 ], 00:37:34.789 "core_count": 1 00:37:34.789 } 00:37:34.789 17:58:42 keyring_linux -- keyring/linux.sh@80 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:37:34.789 17:58:42 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:37:34.789 17:58:42 keyring_linux -- keyring/linux.sh@81 -- # check_keys 0 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@19 -- # local count=0 name= 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:37:34.790 17:58:42 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@22 -- # (( 0 == count )) 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@23 -- # return 00:37:34.790 17:58:42 keyring_linux -- keyring/linux.sh@84 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:37:34.790 17:58:42 keyring_linux -- common/autotest_common.sh@650 -- # local es=0 00:37:34.790 17:58:42 keyring_linux -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:37:34.790 17:58:42 keyring_linux -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:37:34.790 17:58:42 keyring_linux -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:34.790 17:58:42 keyring_linux -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:37:34.790 17:58:42 keyring_linux -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:37:34.790 17:58:42 keyring_linux -- common/autotest_common.sh@653 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:37:34.790 17:58:42 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:37:35.051 [2024-10-17 17:58:42.852400] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:37:35.051 [2024-10-17 17:58:42.853313] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1651360 (107): Transport endpoint is not connected 00:37:35.051 [2024-10-17 17:58:42.854310] nvme_tcp.c:2196:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1651360 (9): Bad file descriptor 00:37:35.051 [2024-10-17 17:58:42.855312] nvme_ctrlr.c:4193:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:37:35.051 [2024-10-17 17:58:42.855319] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:37:35.051 [2024-10-17 17:58:42.855324] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:37:35.051 [2024-10-17 17:58:42.855331] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:37:35.051 request: 00:37:35.051 { 00:37:35.051 "name": "nvme0", 00:37:35.051 "trtype": "tcp", 00:37:35.051 "traddr": "127.0.0.1", 00:37:35.051 "adrfam": "ipv4", 00:37:35.051 "trsvcid": "4420", 00:37:35.051 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:37:35.051 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:37:35.051 "prchk_reftag": false, 00:37:35.051 "prchk_guard": false, 00:37:35.051 "hdgst": false, 00:37:35.051 "ddgst": false, 00:37:35.051 "psk": ":spdk-test:key1", 00:37:35.051 "allow_unrecognized_csi": false, 00:37:35.051 "method": "bdev_nvme_attach_controller", 00:37:35.051 "req_id": 1 00:37:35.051 } 00:37:35.051 Got JSON-RPC error response 00:37:35.051 response: 00:37:35.051 { 00:37:35.051 "code": -5, 00:37:35.051 "message": "Input/output error" 00:37:35.051 } 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@653 -- # es=1 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@1 -- # cleanup 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key0 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@31 -- # local name=key0 sn 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key0 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@33 -- # sn=435014426 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 435014426 00:37:35.051 1 links removed 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key1 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@31 -- # local name=key1 sn 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key1 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key1 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@33 -- # sn=943285474 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 943285474 00:37:35.051 1 links removed 00:37:35.051 17:58:42 keyring_linux -- keyring/linux.sh@41 -- # killprocess 380121 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@950 -- # '[' -z 380121 ']' 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@954 -- # kill -0 380121 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@955 -- # uname 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:35.051 17:58:42 keyring_linux -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 380121 00:37:35.313 17:58:42 keyring_linux -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:37:35.313 17:58:42 keyring_linux -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:37:35.313 17:58:42 keyring_linux -- common/autotest_common.sh@968 -- # echo 'killing process with pid 380121' 00:37:35.313 killing process with pid 380121 00:37:35.313 17:58:42 keyring_linux -- common/autotest_common.sh@969 -- # kill 380121 00:37:35.313 Received shutdown signal, test time was about 1.000000 seconds 00:37:35.313 00:37:35.313 Latency(us) 00:37:35.313 [2024-10-17T15:58:43.232Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:35.313 [2024-10-17T15:58:43.232Z] =================================================================================================================== 00:37:35.313 [2024-10-17T15:58:43.232Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:37:35.313 17:58:42 keyring_linux -- common/autotest_common.sh@974 -- # wait 380121 00:37:35.313 17:58:43 keyring_linux -- keyring/linux.sh@42 -- # killprocess 379963 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@950 -- # '[' -z 379963 ']' 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@954 -- # kill -0 379963 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@955 -- # uname 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 379963 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@968 -- # echo 'killing process with pid 379963' 00:37:35.313 killing process with pid 379963 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@969 -- # kill 379963 00:37:35.313 17:58:43 keyring_linux -- common/autotest_common.sh@974 -- # wait 379963 00:37:35.574 00:37:35.574 real 0m4.058s 00:37:35.574 user 0m7.692s 00:37:35.574 sys 0m1.379s 00:37:35.574 17:58:43 keyring_linux -- common/autotest_common.sh@1126 -- # xtrace_disable 00:37:35.574 17:58:43 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:37:35.574 ************************************ 00:37:35.574 END TEST keyring_linux 00:37:35.574 ************************************ 00:37:35.574 17:58:43 -- spdk/autotest.sh@307 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@320 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@329 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@334 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@351 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:37:35.574 17:58:43 -- spdk/autotest.sh@362 -- # [[ 0 -eq 1 ]] 00:37:35.574 17:58:43 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:37:35.574 17:58:43 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:37:35.574 17:58:43 -- spdk/autotest.sh@374 -- # [[ '' -eq 1 ]] 00:37:35.574 17:58:43 -- spdk/autotest.sh@381 -- # trap - SIGINT SIGTERM EXIT 00:37:35.574 17:58:43 -- spdk/autotest.sh@383 -- # timing_enter post_cleanup 00:37:35.574 17:58:43 -- common/autotest_common.sh@724 -- # xtrace_disable 00:37:35.574 17:58:43 -- common/autotest_common.sh@10 -- # set +x 00:37:35.574 17:58:43 -- spdk/autotest.sh@384 -- # autotest_cleanup 00:37:35.574 17:58:43 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:37:35.574 17:58:43 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:37:35.574 17:58:43 -- common/autotest_common.sh@10 -- # set +x 00:37:43.719 INFO: APP EXITING 00:37:43.719 INFO: killing all VMs 00:37:43.719 INFO: killing vhost app 00:37:43.719 INFO: EXIT DONE 00:37:46.264 0000:80:01.6 (8086 0b00): Already using the ioatdma driver 00:37:46.264 0000:80:01.7 (8086 0b00): Already using the ioatdma driver 00:37:46.264 0000:80:01.4 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:80:01.5 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:80:01.2 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:80:01.3 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:80:01.0 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:80:01.1 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:65:00.0 (144d a80a): Already using the nvme driver 00:37:46.525 0000:00:01.6 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:00:01.7 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:00:01.4 (8086 0b00): Already using the ioatdma driver 00:37:46.525 0000:00:01.5 (8086 0b00): Already using the ioatdma driver 00:37:46.786 0000:00:01.2 (8086 0b00): Already using the ioatdma driver 00:37:46.786 0000:00:01.3 (8086 0b00): Already using the ioatdma driver 00:37:46.786 0000:00:01.0 (8086 0b00): Already using the ioatdma driver 00:37:46.786 0000:00:01.1 (8086 0b00): Already using the ioatdma driver 00:37:50.995 Cleaning 00:37:50.995 Removing: /var/run/dpdk/spdk0/config 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-0 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-1 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-2 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-3 00:37:50.995 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:37:50.995 Removing: /var/run/dpdk/spdk0/hugepage_info 00:37:50.995 Removing: /var/run/dpdk/spdk1/config 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-0 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-1 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-2 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-3 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-0 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-1 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-2 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-3 00:37:50.995 Removing: /var/run/dpdk/spdk1/fbarray_memzone 00:37:50.995 Removing: /var/run/dpdk/spdk1/hugepage_info 00:37:50.995 Removing: /var/run/dpdk/spdk2/config 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-0 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-1 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-2 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-3 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-0 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-1 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-2 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-3 00:37:50.995 Removing: /var/run/dpdk/spdk2/fbarray_memzone 00:37:50.995 Removing: /var/run/dpdk/spdk2/hugepage_info 00:37:50.995 Removing: /var/run/dpdk/spdk3/config 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-0 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-1 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-2 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-3 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-0 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-1 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-2 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-3 00:37:50.995 Removing: /var/run/dpdk/spdk3/fbarray_memzone 00:37:50.995 Removing: /var/run/dpdk/spdk3/hugepage_info 00:37:50.995 Removing: /var/run/dpdk/spdk4/config 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-0 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-1 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-2 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-3 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-0 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-1 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-2 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-3 00:37:50.995 Removing: /var/run/dpdk/spdk4/fbarray_memzone 00:37:50.995 Removing: /var/run/dpdk/spdk4/hugepage_info 00:37:50.995 Removing: /dev/shm/bdev_svc_trace.1 00:37:50.995 Removing: /dev/shm/nvmf_trace.0 00:37:50.995 Removing: /dev/shm/spdk_tgt_trace.pid3998328 00:37:50.996 Removing: /var/run/dpdk/spdk0 00:37:50.996 Removing: /var/run/dpdk/spdk1 00:37:50.996 Removing: /var/run/dpdk/spdk2 00:37:50.996 Removing: /var/run/dpdk/spdk3 00:37:50.996 Removing: /var/run/dpdk/spdk4 00:37:50.996 Removing: /var/run/dpdk/spdk_pid101219 00:37:50.996 Removing: /var/run/dpdk/spdk_pid106033 00:37:50.996 Removing: /var/run/dpdk/spdk_pid113144 00:37:50.996 Removing: /var/run/dpdk/spdk_pid114645 00:37:50.996 Removing: /var/run/dpdk/spdk_pid116296 00:37:50.996 Removing: /var/run/dpdk/spdk_pid117937 00:37:50.996 Removing: /var/run/dpdk/spdk_pid123750 00:37:50.996 Removing: /var/run/dpdk/spdk_pid128721 00:37:50.996 Removing: /var/run/dpdk/spdk_pid138181 00:37:50.996 Removing: /var/run/dpdk/spdk_pid138184 00:37:50.996 Removing: /var/run/dpdk/spdk_pid143277 00:37:50.996 Removing: /var/run/dpdk/spdk_pid143611 00:37:50.996 Removing: /var/run/dpdk/spdk_pid143824 00:37:50.996 Removing: /var/run/dpdk/spdk_pid144282 00:37:50.996 Removing: /var/run/dpdk/spdk_pid144287 00:37:50.996 Removing: /var/run/dpdk/spdk_pid150000 00:37:50.996 Removing: /var/run/dpdk/spdk_pid150650 00:37:50.996 Removing: /var/run/dpdk/spdk_pid156645 00:37:50.996 Removing: /var/run/dpdk/spdk_pid159681 00:37:50.996 Removing: /var/run/dpdk/spdk_pid166423 00:37:50.996 Removing: /var/run/dpdk/spdk_pid173003 00:37:50.996 Removing: /var/run/dpdk/spdk_pid1773 00:37:50.996 Removing: /var/run/dpdk/spdk_pid1777 00:37:50.996 Removing: /var/run/dpdk/spdk_pid183083 00:37:50.996 Removing: /var/run/dpdk/spdk_pid191999 00:37:50.996 Removing: /var/run/dpdk/spdk_pid192002 00:37:50.996 Removing: /var/run/dpdk/spdk_pid19666 00:37:50.996 Removing: /var/run/dpdk/spdk_pid2114 00:37:50.996 Removing: /var/run/dpdk/spdk_pid215540 00:37:50.996 Removing: /var/run/dpdk/spdk_pid216358 00:37:50.996 Removing: /var/run/dpdk/spdk_pid217186 00:37:50.996 Removing: /var/run/dpdk/spdk_pid2176 00:37:50.996 Removing: /var/run/dpdk/spdk_pid217872 00:37:50.996 Removing: /var/run/dpdk/spdk_pid218934 00:37:50.996 Removing: /var/run/dpdk/spdk_pid219628 00:37:50.996 Removing: /var/run/dpdk/spdk_pid220305 00:37:50.996 Removing: /var/run/dpdk/spdk_pid221007 00:37:50.996 Removing: /var/run/dpdk/spdk_pid226331 00:37:50.996 Removing: /var/run/dpdk/spdk_pid226550 00:37:50.996 Removing: /var/run/dpdk/spdk_pid2302 00:37:50.996 Removing: /var/run/dpdk/spdk_pid233816 00:37:50.996 Removing: /var/run/dpdk/spdk_pid234201 00:37:50.996 Removing: /var/run/dpdk/spdk_pid240696 00:37:50.996 Removing: /var/run/dpdk/spdk_pid245761 00:37:50.996 Removing: /var/run/dpdk/spdk_pid258046 00:37:50.996 Removing: /var/run/dpdk/spdk_pid258715 00:37:50.996 Removing: /var/run/dpdk/spdk_pid263801 00:37:50.996 Removing: /var/run/dpdk/spdk_pid264152 00:37:50.996 Removing: /var/run/dpdk/spdk_pid269225 00:37:50.996 Removing: /var/run/dpdk/spdk_pid276128 00:37:50.996 Removing: /var/run/dpdk/spdk_pid279074 00:37:50.996 Removing: /var/run/dpdk/spdk_pid291457 00:37:50.996 Removing: /var/run/dpdk/spdk_pid302127 00:37:50.996 Removing: /var/run/dpdk/spdk_pid304141 00:37:50.996 Removing: /var/run/dpdk/spdk_pid305613 00:37:50.996 Removing: /var/run/dpdk/spdk_pid325307 00:37:50.996 Removing: /var/run/dpdk/spdk_pid330069 00:37:50.996 Removing: /var/run/dpdk/spdk_pid333262 00:37:50.996 Removing: /var/run/dpdk/spdk_pid341041 00:37:50.996 Removing: /var/run/dpdk/spdk_pid341055 00:37:50.996 Removing: /var/run/dpdk/spdk_pid3448 00:37:50.996 Removing: /var/run/dpdk/spdk_pid347166 00:37:50.996 Removing: /var/run/dpdk/spdk_pid349488 00:37:50.996 Removing: /var/run/dpdk/spdk_pid351682 00:37:50.996 Removing: /var/run/dpdk/spdk_pid353174 00:37:50.996 Removing: /var/run/dpdk/spdk_pid355495 00:37:50.996 Removing: /var/run/dpdk/spdk_pid357483 00:37:50.996 Removing: /var/run/dpdk/spdk_pid367478 00:37:50.996 Removing: /var/run/dpdk/spdk_pid368121 00:37:50.996 Removing: /var/run/dpdk/spdk_pid368701 00:37:50.996 Removing: /var/run/dpdk/spdk_pid371518 00:37:50.996 Removing: /var/run/dpdk/spdk_pid372120 00:37:50.996 Removing: /var/run/dpdk/spdk_pid372789 00:37:50.996 Removing: /var/run/dpdk/spdk_pid377681 00:37:50.996 Removing: /var/run/dpdk/spdk_pid377699 00:37:50.996 Removing: /var/run/dpdk/spdk_pid379512 00:37:50.996 Removing: /var/run/dpdk/spdk_pid379963 00:37:50.996 Removing: /var/run/dpdk/spdk_pid380121 00:37:50.996 Removing: /var/run/dpdk/spdk_pid3996833 00:37:50.996 Removing: /var/run/dpdk/spdk_pid3998328 00:37:50.996 Removing: /var/run/dpdk/spdk_pid3999177 00:37:50.996 Removing: /var/run/dpdk/spdk_pid4000216 00:37:50.996 Removing: /var/run/dpdk/spdk_pid4000556 00:37:50.996 Removing: /var/run/dpdk/spdk_pid4001623 00:37:50.996 Removing: /var/run/dpdk/spdk_pid4001806 00:37:50.996 Removing: /var/run/dpdk/spdk_pid4002094 00:37:50.996 Removing: /var/run/dpdk/spdk_pid4003230 00:37:50.996 Removing: /var/run/dpdk/spdk_pid4003882 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4004247 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4004595 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4004937 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4005319 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4005670 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4006021 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4006329 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4007489 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4010875 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4011120 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4011493 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4011701 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4012195 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4012237 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4012881 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4012920 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4013291 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4013484 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4013667 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4013930 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4014442 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4014750 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4015036 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4019755 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4025174 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4037251 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4038165 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4043929 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4044291 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4049697 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4056815 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4059913 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4072508 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4083624 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4085640 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4086666 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4108339 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4113145 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4170553 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4176980 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4184085 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4191469 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4191472 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4192484 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4193487 00:37:51.257 Removing: /var/run/dpdk/spdk_pid4542 00:37:51.257 Removing: /var/run/dpdk/spdk_pid54291 00:37:51.257 Removing: /var/run/dpdk/spdk_pid5641 00:37:51.257 Removing: /var/run/dpdk/spdk_pid59745 00:37:51.257 Removing: /var/run/dpdk/spdk_pid61758 00:37:51.257 Removing: /var/run/dpdk/spdk_pid6287 00:37:51.257 Removing: /var/run/dpdk/spdk_pid6400 00:37:51.257 Removing: /var/run/dpdk/spdk_pid64000 00:37:51.257 Removing: /var/run/dpdk/spdk_pid64169 00:37:51.257 Removing: /var/run/dpdk/spdk_pid64466 00:37:51.257 Removing: /var/run/dpdk/spdk_pid64811 00:37:51.257 Removing: /var/run/dpdk/spdk_pid65530 00:37:51.257 Removing: /var/run/dpdk/spdk_pid6638 00:37:51.257 Removing: /var/run/dpdk/spdk_pid67639 00:37:51.257 Removing: /var/run/dpdk/spdk_pid68836 00:37:51.257 Removing: /var/run/dpdk/spdk_pid69352 00:37:51.257 Removing: /var/run/dpdk/spdk_pid72048 00:37:51.257 Removing: /var/run/dpdk/spdk_pid72758 00:37:51.519 Removing: /var/run/dpdk/spdk_pid73589 00:37:51.519 Removing: /var/run/dpdk/spdk_pid78585 00:37:51.519 Removing: /var/run/dpdk/spdk_pid8475 00:37:51.519 Removing: /var/run/dpdk/spdk_pid85299 00:37:51.519 Removing: /var/run/dpdk/spdk_pid85301 00:37:51.519 Removing: /var/run/dpdk/spdk_pid85303 00:37:51.519 Removing: /var/run/dpdk/spdk_pid90073 00:37:51.519 Removing: /var/run/dpdk/spdk_pid963 00:37:51.519 Removing: /var/run/dpdk/spdk_pid9772 00:37:51.519 Clean 00:37:51.519 17:58:59 -- common/autotest_common.sh@1451 -- # return 0 00:37:51.519 17:58:59 -- spdk/autotest.sh@385 -- # timing_exit post_cleanup 00:37:51.519 17:58:59 -- common/autotest_common.sh@730 -- # xtrace_disable 00:37:51.519 17:58:59 -- common/autotest_common.sh@10 -- # set +x 00:37:51.519 17:58:59 -- spdk/autotest.sh@387 -- # timing_exit autotest 00:37:51.519 17:58:59 -- common/autotest_common.sh@730 -- # xtrace_disable 00:37:51.519 17:58:59 -- common/autotest_common.sh@10 -- # set +x 00:37:51.519 17:58:59 -- spdk/autotest.sh@388 -- # chmod a+r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:37:51.519 17:58:59 -- spdk/autotest.sh@390 -- # [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log ]] 00:37:51.519 17:58:59 -- spdk/autotest.sh@390 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log 00:37:51.519 17:58:59 -- spdk/autotest.sh@392 -- # [[ y == y ]] 00:37:51.519 17:58:59 -- spdk/autotest.sh@394 -- # hostname 00:37:51.519 17:58:59 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -t spdk-cyp-13 -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info 00:37:51.780 geninfo: WARNING: invalid characters removed from testname! 00:38:18.368 17:59:24 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:38:20.282 17:59:28 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/dpdk/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:38:22.829 17:59:30 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:38:24.213 17:59:31 -- spdk/autotest.sh@401 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/examples/vmd/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:38:26.129 17:59:33 -- spdk/autotest.sh@402 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:38:27.573 17:59:35 -- spdk/autotest.sh@403 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:38:29.052 17:59:36 -- spdk/autotest.sh@404 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:38:29.052 17:59:36 -- common/autotest_common.sh@1690 -- $ [[ y == y ]] 00:38:29.052 17:59:36 -- common/autotest_common.sh@1691 -- $ lcov --version 00:38:29.052 17:59:36 -- common/autotest_common.sh@1691 -- $ awk '{print $NF}' 00:38:29.314 17:59:37 -- common/autotest_common.sh@1691 -- $ lt 1.15 2 00:38:29.314 17:59:37 -- scripts/common.sh@373 -- $ cmp_versions 1.15 '<' 2 00:38:29.314 17:59:37 -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:38:29.314 17:59:37 -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:38:29.314 17:59:37 -- scripts/common.sh@336 -- $ IFS=.-: 00:38:29.314 17:59:37 -- scripts/common.sh@336 -- $ read -ra ver1 00:38:29.314 17:59:37 -- scripts/common.sh@337 -- $ IFS=.-: 00:38:29.314 17:59:37 -- scripts/common.sh@337 -- $ read -ra ver2 00:38:29.314 17:59:37 -- scripts/common.sh@338 -- $ local 'op=<' 00:38:29.314 17:59:37 -- scripts/common.sh@340 -- $ ver1_l=2 00:38:29.314 17:59:37 -- scripts/common.sh@341 -- $ ver2_l=1 00:38:29.314 17:59:37 -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:38:29.314 17:59:37 -- scripts/common.sh@344 -- $ case "$op" in 00:38:29.314 17:59:37 -- scripts/common.sh@345 -- $ : 1 00:38:29.314 17:59:37 -- scripts/common.sh@364 -- $ (( v = 0 )) 00:38:29.314 17:59:37 -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:29.314 17:59:37 -- scripts/common.sh@365 -- $ decimal 1 00:38:29.314 17:59:37 -- scripts/common.sh@353 -- $ local d=1 00:38:29.314 17:59:37 -- scripts/common.sh@354 -- $ [[ 1 =~ ^[0-9]+$ ]] 00:38:29.314 17:59:37 -- scripts/common.sh@355 -- $ echo 1 00:38:29.314 17:59:37 -- scripts/common.sh@365 -- $ ver1[v]=1 00:38:29.314 17:59:37 -- scripts/common.sh@366 -- $ decimal 2 00:38:29.314 17:59:37 -- scripts/common.sh@353 -- $ local d=2 00:38:29.314 17:59:37 -- scripts/common.sh@354 -- $ [[ 2 =~ ^[0-9]+$ ]] 00:38:29.314 17:59:37 -- scripts/common.sh@355 -- $ echo 2 00:38:29.314 17:59:37 -- scripts/common.sh@366 -- $ ver2[v]=2 00:38:29.314 17:59:37 -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:38:29.314 17:59:37 -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:38:29.314 17:59:37 -- scripts/common.sh@368 -- $ return 0 00:38:29.314 17:59:37 -- common/autotest_common.sh@1692 -- $ lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:29.314 17:59:37 -- common/autotest_common.sh@1704 -- $ export 'LCOV_OPTS= 00:38:29.314 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:29.314 --rc genhtml_branch_coverage=1 00:38:29.314 --rc genhtml_function_coverage=1 00:38:29.314 --rc genhtml_legend=1 00:38:29.314 --rc geninfo_all_blocks=1 00:38:29.314 --rc geninfo_unexecuted_blocks=1 00:38:29.314 00:38:29.314 ' 00:38:29.314 17:59:37 -- common/autotest_common.sh@1704 -- $ LCOV_OPTS=' 00:38:29.314 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:29.314 --rc genhtml_branch_coverage=1 00:38:29.314 --rc genhtml_function_coverage=1 00:38:29.314 --rc genhtml_legend=1 00:38:29.314 --rc geninfo_all_blocks=1 00:38:29.314 --rc geninfo_unexecuted_blocks=1 00:38:29.314 00:38:29.314 ' 00:38:29.314 17:59:37 -- common/autotest_common.sh@1705 -- $ export 'LCOV=lcov 00:38:29.314 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:29.314 --rc genhtml_branch_coverage=1 00:38:29.314 --rc genhtml_function_coverage=1 00:38:29.314 --rc genhtml_legend=1 00:38:29.314 --rc geninfo_all_blocks=1 00:38:29.314 --rc geninfo_unexecuted_blocks=1 00:38:29.314 00:38:29.314 ' 00:38:29.314 17:59:37 -- common/autotest_common.sh@1705 -- $ LCOV='lcov 00:38:29.314 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:29.314 --rc genhtml_branch_coverage=1 00:38:29.314 --rc genhtml_function_coverage=1 00:38:29.314 --rc genhtml_legend=1 00:38:29.314 --rc geninfo_all_blocks=1 00:38:29.314 --rc geninfo_unexecuted_blocks=1 00:38:29.314 00:38:29.314 ' 00:38:29.314 17:59:37 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:29.314 17:59:37 -- scripts/common.sh@15 -- $ shopt -s extglob 00:38:29.314 17:59:37 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:38:29.314 17:59:37 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:29.314 17:59:37 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:29.314 17:59:37 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:29.314 17:59:37 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:29.314 17:59:37 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:29.314 17:59:37 -- paths/export.sh@5 -- $ export PATH 00:38:29.314 17:59:37 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:29.314 17:59:37 -- common/autobuild_common.sh@485 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:38:29.314 17:59:37 -- common/autobuild_common.sh@486 -- $ date +%s 00:38:29.314 17:59:37 -- common/autobuild_common.sh@486 -- $ mktemp -dt spdk_1729180777.XXXXXX 00:38:29.314 17:59:37 -- common/autobuild_common.sh@486 -- $ SPDK_WORKSPACE=/tmp/spdk_1729180777.NMo7Vk 00:38:29.314 17:59:37 -- common/autobuild_common.sh@488 -- $ [[ -n '' ]] 00:38:29.314 17:59:37 -- common/autobuild_common.sh@492 -- $ '[' -n '' ']' 00:38:29.314 17:59:37 -- common/autobuild_common.sh@495 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:38:29.314 17:59:37 -- common/autobuild_common.sh@499 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:38:29.314 17:59:37 -- common/autobuild_common.sh@501 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:38:29.314 17:59:37 -- common/autobuild_common.sh@502 -- $ get_config_params 00:38:29.314 17:59:37 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:38:29.314 17:59:37 -- common/autotest_common.sh@10 -- $ set +x 00:38:29.314 17:59:37 -- common/autobuild_common.sh@502 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:38:29.314 17:59:37 -- common/autobuild_common.sh@504 -- $ start_monitor_resources 00:38:29.314 17:59:37 -- pm/common@17 -- $ local monitor 00:38:29.314 17:59:37 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:29.314 17:59:37 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:29.314 17:59:37 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:29.314 17:59:37 -- pm/common@21 -- $ date +%s 00:38:29.314 17:59:37 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:29.314 17:59:37 -- pm/common@25 -- $ sleep 1 00:38:29.314 17:59:37 -- pm/common@21 -- $ date +%s 00:38:29.314 17:59:37 -- pm/common@21 -- $ date +%s 00:38:29.314 17:59:37 -- pm/common@21 -- $ date +%s 00:38:29.314 17:59:37 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1729180777 00:38:29.314 17:59:37 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1729180777 00:38:29.314 17:59:37 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1729180777 00:38:29.314 17:59:37 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1729180777 00:38:29.314 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1729180777_collect-cpu-load.pm.log 00:38:29.314 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1729180777_collect-vmstat.pm.log 00:38:29.314 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1729180777_collect-cpu-temp.pm.log 00:38:29.314 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1729180777_collect-bmc-pm.bmc.pm.log 00:38:30.257 17:59:38 -- common/autobuild_common.sh@505 -- $ trap stop_monitor_resources EXIT 00:38:30.257 17:59:38 -- spdk/autopackage.sh@10 -- $ [[ 0 -eq 1 ]] 00:38:30.257 17:59:38 -- spdk/autopackage.sh@14 -- $ timing_finish 00:38:30.257 17:59:38 -- common/autotest_common.sh@736 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:38:30.257 17:59:38 -- common/autotest_common.sh@737 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:38:30.257 17:59:38 -- common/autotest_common.sh@740 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:38:30.257 17:59:38 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:38:30.257 17:59:38 -- pm/common@29 -- $ signal_monitor_resources TERM 00:38:30.257 17:59:38 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:38:30.257 17:59:38 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:30.257 17:59:38 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:38:30.257 17:59:38 -- pm/common@44 -- $ pid=392945 00:38:30.257 17:59:38 -- pm/common@50 -- $ kill -TERM 392945 00:38:30.257 17:59:38 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:30.257 17:59:38 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:38:30.257 17:59:38 -- pm/common@44 -- $ pid=392946 00:38:30.257 17:59:38 -- pm/common@50 -- $ kill -TERM 392946 00:38:30.257 17:59:38 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:30.257 17:59:38 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:38:30.257 17:59:38 -- pm/common@44 -- $ pid=392948 00:38:30.257 17:59:38 -- pm/common@50 -- $ kill -TERM 392948 00:38:30.257 17:59:38 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:38:30.257 17:59:38 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:38:30.257 17:59:38 -- pm/common@44 -- $ pid=392973 00:38:30.257 17:59:38 -- pm/common@50 -- $ sudo -E kill -TERM 392973 00:38:30.518 + [[ -n 3911380 ]] 00:38:30.518 + sudo kill 3911380 00:38:30.529 [Pipeline] } 00:38:30.545 [Pipeline] // stage 00:38:30.550 [Pipeline] } 00:38:30.564 [Pipeline] // timeout 00:38:30.569 [Pipeline] } 00:38:30.584 [Pipeline] // catchError 00:38:30.589 [Pipeline] } 00:38:30.604 [Pipeline] // wrap 00:38:30.610 [Pipeline] } 00:38:30.623 [Pipeline] // catchError 00:38:30.632 [Pipeline] stage 00:38:30.634 [Pipeline] { (Epilogue) 00:38:30.647 [Pipeline] catchError 00:38:30.649 [Pipeline] { 00:38:30.662 [Pipeline] echo 00:38:30.664 Cleanup processes 00:38:30.669 [Pipeline] sh 00:38:30.961 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:38:30.961 393102 /usr/bin/ipmitool sdr dump /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/sdr.cache 00:38:30.961 393642 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:38:30.975 [Pipeline] sh 00:38:31.264 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:38:31.264 ++ grep -v 'sudo pgrep' 00:38:31.264 ++ awk '{print $1}' 00:38:31.264 + sudo kill -9 393102 00:38:31.277 [Pipeline] sh 00:38:31.566 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:38:43.815 [Pipeline] sh 00:38:44.107 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:38:44.107 Artifacts sizes are good 00:38:44.123 [Pipeline] archiveArtifacts 00:38:44.131 Archiving artifacts 00:38:44.263 [Pipeline] sh 00:38:44.552 + sudo chown -R sys_sgci: /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:38:44.568 [Pipeline] cleanWs 00:38:44.578 [WS-CLEANUP] Deleting project workspace... 00:38:44.578 [WS-CLEANUP] Deferred wipeout is used... 00:38:44.586 [WS-CLEANUP] done 00:38:44.587 [Pipeline] } 00:38:44.603 [Pipeline] // catchError 00:38:44.614 [Pipeline] sh 00:38:44.903 + logger -p user.info -t JENKINS-CI 00:38:44.913 [Pipeline] } 00:38:44.926 [Pipeline] // stage 00:38:44.930 [Pipeline] } 00:38:44.943 [Pipeline] // node 00:38:44.948 [Pipeline] End of Pipeline 00:38:44.984 Finished: SUCCESS